[ 493.547682] env[62552]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62552) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 493.548169] env[62552]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62552) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 493.548169] env[62552]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62552) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 493.548423] env[62552]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 493.642513] env[62552]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62552) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 493.652138] env[62552]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62552) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 494.256422] env[62552]: INFO nova.virt.driver [None req-dbbdde8a-9367-45d8-974b-52b49631cb77 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 494.326210] env[62552]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.326360] env[62552]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 494.326400] env[62552]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62552) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 497.457931] env[62552]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-98f415e1-ba39-4b42-b635-3564f15996d5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.474993] env[62552]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62552) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 497.475173] env[62552]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-9a87429e-e86b-476a-ad09-ff95032fa2fd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.508899] env[62552]: INFO oslo_vmware.api [-] Successfully established new session; session ID is aaf92. [ 497.509126] env[62552]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.183s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 497.509698] env[62552]: INFO nova.virt.vmwareapi.driver [None req-dbbdde8a-9367-45d8-974b-52b49631cb77 None None] VMware vCenter version: 7.0.3 [ 497.513133] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c94543f-ab1f-4b6c-9655-f78d6247dbf3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.531230] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3246e79-7c0d-49eb-9b2e-516032d57761 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.537057] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb1daa6-2fcc-4a52-b52a-3365d8df7118 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.543560] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cfd0ce-212e-4a82-ae70-1f6d8e64316f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.556423] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee65fc6-a460-4864-950c-a21f11bc5178 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.562270] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737dc8b9-87e2-4b21-bf68-82e7718bd284 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.591998] env[62552]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-642e7cb5-ffcf-4506-84d0-cd373ec58b73 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.596935] env[62552]: DEBUG nova.virt.vmwareapi.driver [None req-dbbdde8a-9367-45d8-974b-52b49631cb77 None None] Extension org.openstack.compute already exists. {{(pid=62552) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:227}} [ 497.599527] env[62552]: INFO nova.compute.provider_config [None req-dbbdde8a-9367-45d8-974b-52b49631cb77 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 498.106023] env[62552]: DEBUG nova.context [None req-dbbdde8a-9367-45d8-974b-52b49631cb77 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),314587d2-b3d8-4092-8554-fa85da7726aa(cell1) {{(pid=62552) load_cells /opt/stack/nova/nova/context.py:464}} [ 498.106023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.106023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 498.106483] env[62552]: DEBUG oslo_concurrency.lockutils [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 498.107143] env[62552]: DEBUG oslo_concurrency.lockutils [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Acquiring lock "314587d2-b3d8-4092-8554-fa85da7726aa" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.107506] env[62552]: DEBUG oslo_concurrency.lockutils [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Lock "314587d2-b3d8-4092-8554-fa85da7726aa" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 498.108648] env[62552]: DEBUG oslo_concurrency.lockutils [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Lock "314587d2-b3d8-4092-8554-fa85da7726aa" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 498.129249] env[62552]: INFO dbcounter [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Registered counter for database nova_cell0 [ 498.137844] env[62552]: INFO dbcounter [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Registered counter for database nova_cell1 [ 498.141296] env[62552]: DEBUG oslo_db.sqlalchemy.engines [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62552) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 498.142058] env[62552]: DEBUG oslo_db.sqlalchemy.engines [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62552) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 498.146851] env[62552]: ERROR nova.db.main.api [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 498.146851] env[62552]: result = function(*args, **kwargs) [ 498.146851] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 498.146851] env[62552]: return func(*args, **kwargs) [ 498.146851] env[62552]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 498.146851] env[62552]: result = fn(*args, **kwargs) [ 498.146851] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 498.146851] env[62552]: return f(*args, **kwargs) [ 498.146851] env[62552]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 498.146851] env[62552]: return db.service_get_minimum_version(context, binaries) [ 498.146851] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 498.146851] env[62552]: _check_db_access() [ 498.146851] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 498.146851] env[62552]: stacktrace = ''.join(traceback.format_stack()) [ 498.146851] env[62552]: [ 498.148408] env[62552]: ERROR nova.db.main.api [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 498.148408] env[62552]: result = function(*args, **kwargs) [ 498.148408] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 498.148408] env[62552]: return func(*args, **kwargs) [ 498.148408] env[62552]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 498.148408] env[62552]: result = fn(*args, **kwargs) [ 498.148408] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 498.148408] env[62552]: return f(*args, **kwargs) [ 498.148408] env[62552]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 498.148408] env[62552]: return db.service_get_minimum_version(context, binaries) [ 498.148408] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 498.148408] env[62552]: _check_db_access() [ 498.148408] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 498.148408] env[62552]: stacktrace = ''.join(traceback.format_stack()) [ 498.148408] env[62552]: [ 498.152020] env[62552]: WARNING nova.objects.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 498.152020] env[62552]: WARNING nova.objects.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Failed to get minimum service version for cell 314587d2-b3d8-4092-8554-fa85da7726aa [ 498.152020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Acquiring lock "singleton_lock" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 498.152020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Acquired lock "singleton_lock" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 498.152020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Releasing lock "singleton_lock" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 498.152020] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Full set of CONF: {{(pid=62552) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 498.152020] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ******************************************************************************** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 498.152020] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Configuration options gathered from: {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 498.152296] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 498.152296] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 498.152296] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ================================================================================ {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 498.152296] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] allow_resize_to_same_host = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.152296] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] arq_binding_timeout = 300 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.152296] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] backdoor_port = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.152456] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] backdoor_socket = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.152456] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] block_device_allocate_retries = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.152456] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] block_device_allocate_retries_interval = 3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.152770] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cert = self.pem {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.153100] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.153417] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute_monitors = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.153743] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] config_dir = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.154070] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] config_drive_format = iso9660 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.154352] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.154741] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] config_source = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.155069] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] console_host = devstack {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157080] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] control_exchange = nova {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157080] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cpu_allocation_ratio = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157080] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] daemon = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157080] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] debug = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157080] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] default_access_ip_network_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157080] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] default_availability_zone = nova {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157080] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] default_ephemeral_format = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157286] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] default_green_pool_size = 1000 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157286] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157286] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] default_schedule_zone = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157286] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] disk_allocation_ratio = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.157757] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] enable_new_services = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.158102] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] enabled_apis = ['osapi_compute'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.158418] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] enabled_ssl_apis = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.158716] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] flat_injected = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.159033] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] force_config_drive = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.159347] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] force_raw_images = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.159659] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] graceful_shutdown_timeout = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.159963] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] heal_instance_info_cache_interval = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162071] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] host = cpu-1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162071] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162071] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162071] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162071] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162071] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] instance_build_timeout = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162071] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] instance_delete_interval = 300 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162265] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] instance_format = [instance: %(uuid)s] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162265] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] instance_name_template = instance-%08x {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162265] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] instance_usage_audit = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162265] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] instance_usage_audit_period = month {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162587] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.162899] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.163233] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] internal_service_availability_zone = internal {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.163541] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] key = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.163849] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] live_migration_retry_count = 30 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.164168] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] log_color = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.164485] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] log_config_append = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.165031] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167015] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] log_dir = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167015] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] log_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167015] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] log_options = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167015] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] log_rotate_interval = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167015] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] log_rotate_interval_type = days {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167015] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] log_rotation_type = none {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167015] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167261] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167261] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167261] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167261] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167261] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] long_rpc_timeout = 1800 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.167692] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] max_concurrent_builds = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.168023] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] max_concurrent_live_migrations = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.168344] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] max_concurrent_snapshots = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] max_local_block_devices = 3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] max_logfile_count = 30 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] max_logfile_size_mb = 200 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] maximum_instance_delete_attempts = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] metadata_listen = 0.0.0.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] metadata_listen_port = 8775 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] metadata_workers = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171278] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] migrate_max_retries = -1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171278] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] mkisofs_cmd = genisoimage {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171278] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171278] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] my_ip = 10.180.1.21 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171278] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171278] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] network_allocate_retries = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171278] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171449] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171554] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] osapi_compute_listen_port = 8774 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.171830] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] osapi_compute_unique_server_name_scope = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.172041] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] osapi_compute_workers = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.172225] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] password_length = 12 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.172395] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] periodic_enable = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.172560] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] periodic_fuzzy_delay = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.172729] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] pointer_model = usbtablet {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.172901] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] preallocate_images = none {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.173076] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] publish_errors = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.173212] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] pybasedir = /opt/stack/nova {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.173375] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ram_allocation_ratio = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.173540] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] rate_limit_burst = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.173707] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] rate_limit_except_level = CRITICAL {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.173866] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] rate_limit_interval = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.174034] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] reboot_timeout = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.174200] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] reclaim_instance_interval = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.174357] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] record = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.174532] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] reimage_timeout_per_gb = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.174711] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] report_interval = 120 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.174868] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] rescue_timeout = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.175054] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] reserved_host_cpus = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.175232] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] reserved_host_disk_mb = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.175396] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] reserved_host_memory_mb = 512 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.175558] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] reserved_huge_pages = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.175720] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] resize_confirm_window = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.175878] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] resize_fs_using_block_device = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.176045] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] resume_guests_state_on_host_boot = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.176215] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.176377] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] rpc_response_timeout = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.176532] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] run_external_periodic_tasks = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.176698] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] running_deleted_instance_action = reap {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.176857] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.177024] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] running_deleted_instance_timeout = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.177226] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler_instance_sync_interval = 120 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.177405] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_down_time = 720 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.177577] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] servicegroup_driver = db {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.177734] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] shell_completion = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.177896] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] shelved_offload_time = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.178069] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] shelved_poll_interval = 3600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.178277] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] shutdown_timeout = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.178464] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] source_is_ipv6 = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.178626] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ssl_only = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.178872] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.179051] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] sync_power_state_interval = 600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.179218] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] sync_power_state_pool_size = 1000 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.179546] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] syslog_log_facility = LOG_USER {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.179605] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] tempdir = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.179729] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] timeout_nbd = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.179908] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] transport_url = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.180061] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] update_resources_interval = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.180223] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] use_cow_images = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.180387] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] use_eventlog = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.180539] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] use_journal = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.180695] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] use_json = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.180851] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] use_rootwrap_daemon = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.181023] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] use_stderr = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.181177] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] use_syslog = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.181332] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vcpu_pin_set = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.181495] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plugging_is_fatal = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.181671] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plugging_timeout = 300 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.181835] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] virt_mkfs = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.181993] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] volume_usage_poll_interval = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.182164] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] watch_log_file = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.182330] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] web = /usr/share/spice-html5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 498.182510] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.182676] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.182838] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.183015] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_concurrency.disable_process_locking = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.183308] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.183491] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.183654] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.183822] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.183988] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.184167] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.184348] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.auth_strategy = keystone {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.184517] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.compute_link_prefix = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.184693] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.184870] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.dhcp_domain = novalocal {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.185050] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.enable_instance_password = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.185219] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.glance_link_prefix = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.185386] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.185556] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.185758] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.instance_list_per_project_cells = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.185963] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.list_records_by_skipping_down_cells = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.186166] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.local_metadata_per_cell = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.186345] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.max_limit = 1000 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.186513] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.metadata_cache_expiration = 15 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.186684] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.neutron_default_tenant_id = default {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.186860] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.response_validation = warn {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.187037] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.use_neutron_default_nets = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.187239] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.187413] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.187584] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.187758] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.187929] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.vendordata_dynamic_targets = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.188115] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.vendordata_jsonfile_path = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.188292] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.188479] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.backend = dogpile.cache.memcached {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.188650] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.backend_argument = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.188821] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.config_prefix = cache.oslo {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.188990] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.dead_timeout = 60.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.189168] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.debug_cache_backend = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.189331] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.enable_retry_client = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.189508] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.enable_socket_keepalive = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.189688] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.enabled = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.189852] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.enforce_fips_mode = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.190025] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.expiration_time = 600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.190194] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.hashclient_retry_attempts = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.190363] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.190527] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_dead_retry = 300 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.190686] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_password = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.190849] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.191014] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.191182] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_pool_maxsize = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.191347] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.191508] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_sasl_enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.191687] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.191872] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.192015] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.memcache_username = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.192186] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.proxies = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.192352] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.redis_db = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.192514] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.redis_password = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.192683] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.192858] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.193036] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.redis_server = localhost:6379 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.193204] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.redis_socket_timeout = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.193364] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.redis_username = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.193529] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.retry_attempts = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.193753] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.retry_delay = 0.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.193930] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.socket_keepalive_count = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.194108] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.socket_keepalive_idle = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.194276] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.socket_keepalive_interval = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.194434] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.tls_allowed_ciphers = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.194590] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.tls_cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.194744] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.tls_certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.194912] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.tls_enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.195074] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cache.tls_keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.195246] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.195422] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.auth_type = password {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.195584] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.195758] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.195918] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.196094] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.196259] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.cross_az_attach = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.196420] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.debug = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.196579] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.endpoint_template = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.196743] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.http_retries = 3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.196904] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.197071] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.197270] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.os_region_name = RegionOne {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.197444] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.197605] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cinder.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.197776] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.197937] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.cpu_dedicated_set = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.198153] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.cpu_shared_set = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.198307] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.image_type_exclude_list = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.198473] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.198641] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.198803] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.198964] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.199147] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.199315] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.resource_provider_association_refresh = 300 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.199505] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.199687] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.shutdown_retry_interval = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.199870] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.200062] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] conductor.workers = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.200245] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] console.allowed_origins = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.200408] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] console.ssl_ciphers = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.200578] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] console.ssl_minimum_version = default {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.200745] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] consoleauth.enforce_session_timeout = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.200913] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] consoleauth.token_ttl = 600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.201098] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.201262] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.201426] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.201586] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.connect_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.201747] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.connect_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.201906] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.endpoint_override = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.202082] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.202244] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.202405] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.max_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.202565] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.min_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.202722] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.region_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.202882] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.retriable_status_codes = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.203049] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.service_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.203229] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.service_type = accelerator {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.203384] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.203545] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.status_code_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.203703] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.status_code_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.203861] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.204052] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.204219] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] cyborg.version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.204400] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.backend = sqlalchemy {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.204569] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.connection = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.204736] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.connection_debug = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.204904] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.connection_parameters = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.205085] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.connection_recycle_time = 3600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.205249] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.connection_trace = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.205413] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.db_inc_retry_interval = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.205577] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.db_max_retries = 20 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.205739] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.db_max_retry_interval = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.205902] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.db_retry_interval = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.206078] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.max_overflow = 50 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.206243] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.max_pool_size = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.206408] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.max_retries = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.206576] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.206735] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.mysql_wsrep_sync_wait = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.206894] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.pool_timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.207070] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.retry_interval = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.207257] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.slave_connection = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.207428] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.sqlite_synchronous = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.207592] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] database.use_db_reconnect = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.207772] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.backend = sqlalchemy {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.207940] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.connection = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.208135] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.connection_debug = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.208321] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.connection_parameters = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.208488] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.connection_recycle_time = 3600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.208651] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.connection_trace = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.208812] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.db_inc_retry_interval = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.208977] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.db_max_retries = 20 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.209159] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.db_max_retry_interval = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.209324] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.db_retry_interval = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.209507] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.max_overflow = 50 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.209683] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.max_pool_size = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.209848] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.max_retries = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.210026] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.210194] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.210357] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.pool_timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.210523] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.retry_interval = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.210682] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.slave_connection = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.210870] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] api_database.sqlite_synchronous = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.211077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] devices.enabled_mdev_types = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.211263] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.211438] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.211602] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ephemeral_storage_encryption.enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.211769] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.211942] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.api_servers = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.212120] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.212288] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.212454] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.212616] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.connect_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.212773] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.connect_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.212935] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.debug = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.213115] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.default_trusted_certificate_ids = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.213282] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.enable_certificate_validation = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.213446] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.enable_rbd_download = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.213606] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.endpoint_override = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.213772] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.213931] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.214105] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.max_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.214269] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.min_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.214435] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.num_retries = 3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.214604] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.rbd_ceph_conf = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.214769] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.rbd_connect_timeout = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.214936] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.rbd_pool = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.215159] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.rbd_user = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.215274] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.region_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.215438] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.retriable_status_codes = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.215595] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.service_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.215760] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.service_type = image {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.215928] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.216101] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.status_code_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.216267] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.status_code_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.216430] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.216611] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.216775] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.verify_glance_signatures = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.216936] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] glance.version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.217142] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] guestfs.debug = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.217318] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.217485] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.auth_type = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.217643] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.217802] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.217962] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.218157] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.connect_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.218333] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.connect_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.218493] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.endpoint_override = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.218654] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.218813] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.218973] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.max_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.219146] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.min_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.219305] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.region_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.219485] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.retriable_status_codes = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.219658] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.service_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.219835] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.service_type = shared-file-system {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.220007] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.share_apply_policy_timeout = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.220179] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.220342] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.status_code_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.220500] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.status_code_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.220659] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.220835] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.220996] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] manila.version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.221182] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] mks.enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.221545] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.221749] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] image_cache.manager_interval = 2400 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.221923] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] image_cache.precache_concurrency = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.222109] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] image_cache.remove_unused_base_images = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.222282] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.222451] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.222630] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] image_cache.subdirectory_name = _base {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.222808] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.api_max_retries = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.222974] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.api_retry_interval = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.223148] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.223312] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.auth_type = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.223470] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.223626] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.223789] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.223951] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.conductor_group = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.224126] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.connect_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.224290] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.connect_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.224449] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.endpoint_override = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.224610] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.224768] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.224928] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.max_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.225102] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.min_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.225274] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.peer_list = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.225427] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.region_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.225587] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.retriable_status_codes = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.225751] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.serial_console_state_timeout = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.225908] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.service_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.226088] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.service_type = baremetal {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.226248] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.shard = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.226411] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.226571] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.status_code_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.226728] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.status_code_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.226887] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.227078] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.227270] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ironic.version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.227459] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.227635] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] key_manager.fixed_key = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.227819] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.227982] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.barbican_api_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.228173] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.barbican_endpoint = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.228362] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.barbican_endpoint_type = public {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.228524] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.barbican_region_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.228682] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.228839] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.229009] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.229180] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.229339] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.229529] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.number_of_retries = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.229706] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.retry_delay = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.229873] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.send_service_user_token = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.230046] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.230215] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.230380] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.verify_ssl = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.230548] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican.verify_ssl_path = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.230721] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican_service_user.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.230886] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican_service_user.auth_type = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.231057] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican_service_user.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.231218] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican_service_user.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.231383] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican_service_user.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.231545] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican_service_user.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.231701] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican_service_user.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.231863] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican_service_user.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.232030] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] barbican_service_user.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.232202] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.approle_role_id = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.232363] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.approle_secret_id = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.232528] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.kv_mountpoint = secret {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.232687] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.kv_path = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.232851] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.kv_version = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.233024] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.namespace = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.233179] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.root_token_id = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.233339] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.ssl_ca_crt_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.233506] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.timeout = 60.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.233666] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.use_ssl = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.233831] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.234007] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.234179] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.auth_type = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.234342] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.234499] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.234662] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.234823] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.connect_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.234984] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.connect_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.235155] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.endpoint_override = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.235317] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.235474] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.235630] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.max_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.235785] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.min_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.235943] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.region_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.236116] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.retriable_status_codes = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.236279] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.service_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.236450] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.service_type = identity {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.236612] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.236774] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.status_code_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.236931] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.status_code_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.237134] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.237315] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.237484] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] keystone.version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.237682] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.connection_uri = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.237843] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.cpu_mode = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.238045] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.238246] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.cpu_models = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.238428] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.cpu_power_governor_high = performance {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.238595] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.238761] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.cpu_power_management = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.238933] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.239113] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.device_detach_attempts = 8 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.239284] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.device_detach_timeout = 20 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.239454] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.disk_cachemodes = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.239612] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.disk_prefix = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.239782] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.enabled_perf_events = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.239950] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.file_backed_memory = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.240133] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.gid_maps = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.240298] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.hw_disk_discard = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.240460] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.hw_machine_type = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.240629] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.images_rbd_ceph_conf = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.240793] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.240955] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.241137] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.images_rbd_glance_store_name = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.241309] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.images_rbd_pool = rbd {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.241479] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.images_type = default {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.241639] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.images_volume_group = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.241800] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.inject_key = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.241963] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.inject_partition = -2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.242138] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.inject_password = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.242304] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.iscsi_iface = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.242467] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.iser_use_multipath = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.242630] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.242791] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.242952] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_downtime = 500 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.243127] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.243291] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.243450] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_inbound_addr = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.243610] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.243770] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.243926] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_scheme = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.244107] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_timeout_action = abort {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.244276] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_tunnelled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.244433] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_uri = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.244592] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.live_migration_with_native_tls = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.244751] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.max_queues = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.244913] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.245167] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.245333] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.nfs_mount_options = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.245636] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.245807] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.245971] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.246145] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.246311] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.246475] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.num_pcie_ports = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.246640] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.246805] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.pmem_namespaces = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.246963] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.quobyte_client_cfg = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.247289] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.247471] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.247639] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.247805] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.247965] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rbd_secret_uuid = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.248158] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rbd_user = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.248340] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.248511] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.248671] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rescue_image_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.248829] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rescue_kernel_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.248986] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rescue_ramdisk_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.249168] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.249331] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.rx_queue_size = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.249525] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.smbfs_mount_options = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.249808] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.249983] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.snapshot_compression = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.250163] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.snapshot_image_format = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.250386] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.250554] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.sparse_logical_volumes = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.250718] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.swtpm_enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.250885] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.swtpm_group = tss {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.251062] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.swtpm_user = tss {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.251237] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.sysinfo_serial = unique {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.251399] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.tb_cache_size = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.251559] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.tx_queue_size = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.251725] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.uid_maps = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.251887] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.use_virtio_for_bridges = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.252066] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.virt_type = kvm {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.252239] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.volume_clear = zero {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.252404] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.volume_clear_size = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.252572] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.volume_use_multipath = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.252730] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.vzstorage_cache_path = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.252897] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.253073] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.253244] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.253412] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.253680] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.253854] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.vzstorage_mount_user = stack {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.254029] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.254207] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.254379] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.auth_type = password {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.254542] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.254703] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.254868] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.255039] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.connect_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.255206] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.connect_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.255376] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.default_floating_pool = public {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.255546] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.endpoint_override = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.255705] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.extension_sync_interval = 600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.255867] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.http_retries = 3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.256037] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.256203] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.256366] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.max_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.256534] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.256693] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.min_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.256859] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.ovs_bridge = br-int {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.257033] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.physnets = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.257240] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.region_name = RegionOne {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.257409] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.retriable_status_codes = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.257580] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.service_metadata_proxy = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.257738] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.service_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.257904] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.service_type = network {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.258077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.258264] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.status_code_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.258428] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.status_code_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.258590] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.258770] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.258938] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] neutron.version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.259124] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] notifications.bdms_in_notifications = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.259309] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] notifications.default_level = INFO {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.259504] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] notifications.notification_format = unversioned {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.259682] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] notifications.notify_on_state_change = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.259862] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.260051] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] pci.alias = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.260226] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] pci.device_spec = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.260449] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] pci.report_in_placement = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.260750] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.261052] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.auth_type = password {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.261344] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.261619] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.261819] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.261997] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.262184] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.connect_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.262351] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.connect_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.262513] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.default_domain_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.262677] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.default_domain_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.262836] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.domain_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.262997] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.domain_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.263172] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.endpoint_override = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.263338] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.263498] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.263658] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.max_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.263815] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.min_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.263982] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.password = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.264157] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.project_domain_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.264329] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.project_domain_name = Default {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.264493] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.project_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.264664] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.project_name = service {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.264833] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.region_name = RegionOne {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.264994] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.retriable_status_codes = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.265170] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.service_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.265339] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.service_type = placement {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.265514] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.265656] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.status_code_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.265815] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.status_code_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.265973] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.system_scope = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.266145] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.266306] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.trust_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.266465] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.user_domain_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.266631] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.user_domain_name = Default {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.266789] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.user_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.266957] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.username = nova {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.267188] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.267362] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] placement.version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.267551] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.cores = 20 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.267767] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.count_usage_from_placement = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.267949] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.268156] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.injected_file_content_bytes = 10240 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.268343] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.injected_file_path_length = 255 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.268519] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.injected_files = 5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.268687] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.instances = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.268854] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.key_pairs = 100 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.269028] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.metadata_items = 128 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.269201] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.ram = 51200 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.269365] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.recheck_quota = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.269532] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.server_group_members = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.269698] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] quota.server_groups = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.269870] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.270042] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.270230] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.image_metadata_prefilter = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.270444] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.270620] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.max_attempts = 3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.270843] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.max_placement_results = 1000 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.271033] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.271205] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.271371] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.271547] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] scheduler.workers = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.271721] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.271891] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.272085] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.272260] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.272428] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.272590] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.272753] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.272939] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.273122] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.host_subset_size = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.273291] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.273450] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.273616] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.273781] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.isolated_hosts = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.273945] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.isolated_images = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.274122] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.274288] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.274451] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.274611] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.pci_in_placement = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.274775] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.274933] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.275107] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.275286] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.275449] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.275610] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.275769] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.track_instance_changes = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.275945] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.276124] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] metrics.required = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.276290] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] metrics.weight_multiplier = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.276451] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.276612] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] metrics.weight_setting = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.276924] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.277179] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] serial_console.enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.277383] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] serial_console.port_range = 10000:20000 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.277559] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.277727] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.277895] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] serial_console.serialproxy_port = 6083 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.278078] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.278256] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.auth_type = password {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.278418] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.278578] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.278741] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.278903] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.279070] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.279250] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.send_service_user_token = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.279416] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.279574] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] service_user.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.279742] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.agent_enabled = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.279902] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.280221] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.280416] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.280586] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.html5proxy_port = 6082 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.280748] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.image_compression = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.280908] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.jpeg_compression = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.281077] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.playback_compression = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.281246] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.require_secure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.281414] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.server_listen = 127.0.0.1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.281580] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.281735] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.streaming_mode = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.281892] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] spice.zlib_compression = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.282066] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] upgrade_levels.baseapi = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.282239] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] upgrade_levels.compute = auto {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.282399] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] upgrade_levels.conductor = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.282553] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] upgrade_levels.scheduler = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.282717] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.282880] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.283044] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vendordata_dynamic_auth.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.283204] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vendordata_dynamic_auth.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.283366] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.283528] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vendordata_dynamic_auth.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.283684] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.283843] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.284008] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vendordata_dynamic_auth.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.284197] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.api_retry_count = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.284359] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.ca_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.284528] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.284694] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.cluster_name = testcl1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.284859] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.connection_pool_size = 10 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.285030] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.console_delay_seconds = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.285204] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.datastore_regex = ^datastore.* {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.285405] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.285577] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.host_password = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.285741] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.host_port = 443 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.285907] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.host_username = administrator@vsphere.local {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.286089] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.insecure = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.286255] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.integration_bridge = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.286418] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.maximum_objects = 100 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.286578] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.pbm_default_policy = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.286739] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.pbm_enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.286895] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.pbm_wsdl_location = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.287073] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.287244] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.serial_port_proxy_uri = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.287405] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.serial_port_service_uri = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.287571] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.task_poll_interval = 0.5 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.287741] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.use_linked_clone = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.287907] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.vnc_keymap = en-us {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.288084] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.vnc_port = 5900 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.288280] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vmware.vnc_port_total = 10000 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.288469] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.auth_schemes = ['none'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.288645] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.288936] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.289132] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.289306] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.novncproxy_port = 6080 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.289481] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.server_listen = 127.0.0.1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.289650] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.289808] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.vencrypt_ca_certs = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.289966] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.vencrypt_client_cert = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.290140] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vnc.vencrypt_client_key = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.290319] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.290482] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.disable_deep_image_inspection = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.290640] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.290799] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.290959] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.291129] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.disable_rootwrap = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.291303] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.enable_numa_live_migration = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.291469] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.291628] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.291787] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.291947] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.libvirt_disable_apic = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.292119] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.292284] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.292445] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.292604] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.292760] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.292931] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.293088] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.293250] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.293409] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.293575] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.293755] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.293924] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.client_socket_timeout = 900 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.294108] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.default_pool_size = 1000 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.294281] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.keep_alive = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.294448] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.max_header_line = 16384 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.294609] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.294769] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.ssl_ca_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.294930] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.ssl_cert_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.295102] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.ssl_key_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.295268] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.tcp_keepidle = 600 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.295445] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.295609] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] zvm.ca_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.295767] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] zvm.cloud_connector_url = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.296055] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.296236] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] zvm.reachable_timeout = 300 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.296420] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.enforce_new_defaults = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.296792] env[62552]: WARNING oslo_config.cfg [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 498.296977] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.enforce_scope = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.297195] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.policy_default_rule = default {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.297393] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.297572] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.policy_file = policy.yaml {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.297746] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.297910] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.298089] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.298283] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.298454] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.298626] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.298803] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.298980] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.connection_string = messaging:// {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.299162] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.enabled = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.299335] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.es_doc_type = notification {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.299501] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.es_scroll_size = 10000 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.299670] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.es_scroll_time = 2m {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.299832] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.filter_error_trace = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.300007] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.hmac_keys = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.300184] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.sentinel_service_name = mymaster {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.300354] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.socket_timeout = 0.1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.300516] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.trace_requests = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.300674] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler.trace_sqlalchemy = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.300858] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler_jaeger.process_tags = {} {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.301031] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler_jaeger.service_name_prefix = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.301198] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] profiler_otlp.service_name_prefix = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.301364] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] remote_debug.host = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.301524] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] remote_debug.port = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.301704] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.301867] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.302041] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.302208] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.302371] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.302528] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.302685] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.302845] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.303041] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.303188] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.303347] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.303518] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.303685] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.303850] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.304026] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.304198] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.304363] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.304535] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.304698] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.304859] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.305033] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.305232] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.305435] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.305690] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.305943] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.306275] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.306620] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.306823] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.307033] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.307236] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.ssl = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.307447] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.307684] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.307942] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.308229] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.308462] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.308672] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.308932] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.309181] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_notifications.retry = -1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.309441] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.309696] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.309937] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.auth_section = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.310174] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.auth_type = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.310399] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.cafile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.310609] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.certfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.310836] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.collect_timing = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.311042] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.connect_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.311310] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.connect_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.311590] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.endpoint_id = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.311869] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.endpoint_override = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.312127] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.insecure = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.312400] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.keyfile = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.312647] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.max_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.312931] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.min_version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.313191] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.region_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.313471] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.retriable_status_codes = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.313765] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.service_name = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.313999] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.service_type = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.314580] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.split_loggers = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.314580] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.status_code_retries = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.314800] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.status_code_retry_delay = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.315048] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.timeout = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.315354] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.valid_interfaces = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.315631] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_limit.version = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.315904] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_reports.file_event_handler = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.317535] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.317535] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] oslo_reports.log_dir = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.317535] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.317535] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.317535] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.317535] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.317782] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.318018] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.318274] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.318533] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_ovs_privileged.group = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.319044] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.319044] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.319278] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.319504] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] vif_plug_ovs_privileged.user = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.319775] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.320036] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.320412] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.320550] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.320822] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.322607] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.322607] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.322607] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.322607] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.322607] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_ovs.isolate_vif = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.322607] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.322807] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.322904] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.323256] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.323512] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] os_vif_ovs.per_port_bridge = False {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.323707] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] privsep_osbrick.capabilities = [21] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.323982] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] privsep_osbrick.group = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.324224] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] privsep_osbrick.helper_command = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.324490] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.324721] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.324951] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] privsep_osbrick.user = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.325236] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.325499] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] nova_sys_admin.group = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.325737] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] nova_sys_admin.helper_command = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.325964] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.326245] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.327579] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] nova_sys_admin.user = None {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 498.327579] env[62552]: DEBUG oslo_service.service [None req-65c65766-fd24-4fa9-913f-694ac2deaf03 None None] ******************************************************************************** {{(pid=62552) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 498.327579] env[62552]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 498.831372] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Getting list of instances from cluster (obj){ [ 498.831372] env[62552]: value = "domain-c8" [ 498.831372] env[62552]: _type = "ClusterComputeResource" [ 498.831372] env[62552]: } {{(pid=62552) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 498.832532] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50b80ab-b51d-4595-9c6c-66b86fc2dc79 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.841154] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Got total of 0 instances {{(pid=62552) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 498.841667] env[62552]: WARNING nova.virt.vmwareapi.driver [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 498.842132] env[62552]: INFO nova.virt.node [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Generated node identity b52e1b51-8a99-4529-bad8-00cd27fc0fb8 [ 498.842362] env[62552]: INFO nova.virt.node [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Wrote node identity b52e1b51-8a99-4529-bad8-00cd27fc0fb8 to /opt/stack/data/n-cpu-1/compute_id [ 499.344833] env[62552]: WARNING nova.compute.manager [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Compute nodes ['b52e1b51-8a99-4529-bad8-00cd27fc0fb8'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 500.350917] env[62552]: INFO nova.compute.manager [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 501.356036] env[62552]: WARNING nova.compute.manager [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 501.356401] env[62552]: DEBUG oslo_concurrency.lockutils [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 501.356532] env[62552]: DEBUG oslo_concurrency.lockutils [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 501.356683] env[62552]: DEBUG oslo_concurrency.lockutils [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 501.356839] env[62552]: DEBUG nova.compute.resource_tracker [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 501.357843] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c840c2b3-e9f0-42a8-bbb4-a6b03acb466f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.365611] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e81e2c9-181a-44d5-aeb2-cc26086bfd19 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.378930] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5d96ad-c3b9-4dad-b3d6-441acc69fa8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.385146] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-892ed1be-ce3b-4aca-aee6-c1aacbbd2a76 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.413803] env[62552]: DEBUG nova.compute.resource_tracker [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181541MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 501.413959] env[62552]: DEBUG oslo_concurrency.lockutils [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 501.414141] env[62552]: DEBUG oslo_concurrency.lockutils [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 501.917140] env[62552]: WARNING nova.compute.resource_tracker [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] No compute node record for cpu-1:b52e1b51-8a99-4529-bad8-00cd27fc0fb8: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host b52e1b51-8a99-4529-bad8-00cd27fc0fb8 could not be found. [ 502.420949] env[62552]: INFO nova.compute.resource_tracker [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 [ 503.927862] env[62552]: DEBUG nova.compute.resource_tracker [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 503.928177] env[62552]: DEBUG nova.compute.resource_tracker [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 504.177790] env[62552]: INFO nova.scheduler.client.report [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] [req-45593e5d-04ab-445e-b510-1e27a9158443] Created resource provider record via placement API for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 504.213513] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4023420c-7d2a-48f2-bca1-e8f328eebba7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.221920] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3712490e-0a8e-45dd-bc00-bfa617bcdd9e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.255837] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db56fc2-ae8d-4b1b-8e55-6993e0e3988b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.266997] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1683b2c-36c9-4f60-ba62-66e67c3961ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.280448] env[62552]: DEBUG nova.compute.provider_tree [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 504.821273] env[62552]: DEBUG nova.scheduler.client.report [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 504.821517] env[62552]: DEBUG nova.compute.provider_tree [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 0 to 1 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 504.821835] env[62552]: DEBUG nova.compute.provider_tree [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 504.868999] env[62552]: DEBUG nova.compute.provider_tree [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 1 to 2 during operation: update_traits {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 505.374194] env[62552]: DEBUG nova.compute.resource_tracker [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 505.374457] env[62552]: DEBUG oslo_concurrency.lockutils [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.960s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 505.374574] env[62552]: DEBUG nova.service [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Creating RPC server for service compute {{(pid=62552) start /opt/stack/nova/nova/service.py:186}} [ 505.387955] env[62552]: DEBUG nova.service [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] Join ServiceGroup membership for this service compute {{(pid=62552) start /opt/stack/nova/nova/service.py:203}} [ 505.388182] env[62552]: DEBUG nova.servicegroup.drivers.db [None req-84d87157-fe87-4e9a-85e3-495f28ae71cd None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62552) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 535.390805] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 535.896113] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Getting list of instances from cluster (obj){ [ 535.896113] env[62552]: value = "domain-c8" [ 535.896113] env[62552]: _type = "ClusterComputeResource" [ 535.896113] env[62552]: } {{(pid=62552) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 535.896113] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb7416c-0652-4c8f-bb07-abbc07bb8968 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.907014] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Got total of 0 instances {{(pid=62552) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 535.907316] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 535.907707] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Getting list of instances from cluster (obj){ [ 535.907707] env[62552]: value = "domain-c8" [ 535.907707] env[62552]: _type = "ClusterComputeResource" [ 535.907707] env[62552]: } {{(pid=62552) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 535.908893] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157ba2b5-b867-4092-9a09-d09532543a17 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.918268] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Got total of 0 instances {{(pid=62552) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 540.017715] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "cffa534e-d51e-484c-b0e4-b27f6d094203" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.017715] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "cffa534e-d51e-484c-b0e4-b27f6d094203" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.519694] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 541.056372] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.058026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.059162] env[62552]: INFO nova.compute.claims [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.486937] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "373f9654-0871-4957-abaa-d4724a9c43c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.487660] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "373f9654-0871-4957-abaa-d4724a9c43c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.991767] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 542.149496] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6299133c-706a-47ba-8761-dcba25f63839 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.164806] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7143bc27-009a-4e7a-bcd8-dd468169dc89 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.207614] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d2bdfb-0d60-4bdc-8c7a-80ae44cf6f8a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.215559] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e839e8-d3f8-4bba-be8c-3eb33278802b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.238432] env[62552]: DEBUG nova.compute.provider_tree [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.522908] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.741616] env[62552]: DEBUG nova.scheduler.client.report [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 542.977131] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquiring lock "6240d4e9-79ff-4c84-9d04-49879088bde4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.977131] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "6240d4e9-79ff-4c84-9d04-49879088bde4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.057843] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Acquiring lock "c56252ef-675a-4e4e-86ee-6cf36157f04d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.057843] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Lock "c56252ef-675a-4e4e-86ee-6cf36157f04d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.250606] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.191s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.250606] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 543.252218] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.729s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.254553] env[62552]: INFO nova.compute.claims [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 543.480787] env[62552]: DEBUG nova.compute.manager [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 543.560580] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 543.685262] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "2f52f72d-176a-4278-9592-b06ee4807df1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.686154] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "2f52f72d-176a-4278-9592-b06ee4807df1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.764687] env[62552]: DEBUG nova.compute.utils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 543.766615] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 543.768208] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 544.005751] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.086062] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.143604] env[62552]: DEBUG nova.policy [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae73aca55f7b491a8cd7bd17750dc623', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf213723411c477ea325512314f7aeb0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.191128] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 544.276917] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 544.387014] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c1fdab-7a4e-4c27-bffd-39bab70c20d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.394274] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34cb3fcf-74a9-434d-a42b-65f19e9b4382 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.425342] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c4b5c8-4a00-4db2-8872-68fb7027153f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.432535] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd7935a-ec52-4196-8cdc-d33df8efc5d1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.449129] env[62552]: DEBUG nova.compute.provider_tree [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.723604] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.953692] env[62552]: DEBUG nova.scheduler.client.report [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 544.982110] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Successfully created port: 3da47864-9e50-4957-a800-52e48d0d24b4 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 545.291289] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 545.322547] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 545.322796] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 545.322952] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.323241] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 545.323308] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.323422] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 545.323638] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 545.324096] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 545.324224] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 545.324403] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 545.324564] env[62552]: DEBUG nova.virt.hardware [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 545.325504] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca849c1-5981-4629-8463-3820a41e7c03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.334693] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff3770a-8eae-40be-8489-01ce83a78ada {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.350699] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12008f7b-4286-4d8f-835f-65ed2381c52f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.461695] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.207s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.461695] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 545.464486] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.459s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.466070] env[62552]: INFO nova.compute.claims [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 545.665175] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Acquiring lock "05687d2d-0760-4c5d-a231-4ef7c2d440e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.665484] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Lock "05687d2d-0760-4c5d-a231-4ef7c2d440e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.975971] env[62552]: DEBUG nova.compute.utils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 545.978530] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 545.978811] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 546.108228] env[62552]: DEBUG nova.policy [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '384d6877b7384d21b12a155d13056a7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f13a11803d1b44178f053cc0a4affedd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 546.168138] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 546.313480] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Acquiring lock "b3dcf7cf-5807-4483-8ca0-a245931c7dc3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.313726] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Lock "b3dcf7cf-5807-4483-8ca0-a245931c7dc3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.490602] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 546.633129] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c9da4f-540c-443d-9918-564ad2a01466 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.642923] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93f7042-5864-4f02-9ebf-c72f2a6aac0d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.680721] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe4b976-fb93-4a20-9a64-c0b31c5720f8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.692129] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59253d9b-c8b8-490c-9b42-37ef5f7c55c7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.708499] env[62552]: DEBUG nova.compute.provider_tree [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.711402] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.753031] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Successfully created port: 3175847b-1a90-4540-8a21-c434f25d077c {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.816537] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 547.211776] env[62552]: DEBUG nova.scheduler.client.report [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 547.346820] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.510175] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 547.541050] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 547.541575] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 547.541575] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 547.541575] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 547.541818] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 547.541997] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 547.542391] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 547.542391] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 547.542519] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 547.542676] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 547.542842] env[62552]: DEBUG nova.virt.hardware [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 547.543734] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf92013-5d81-4897-85d0-5760edd5c1f0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.556020] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf9ec99-f36c-4c77-93c6-6c6f4c0cf770 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.718948] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.254s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.719435] env[62552]: DEBUG nova.compute.manager [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 547.729099] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.640s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.730787] env[62552]: INFO nova.compute.claims [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.800556] env[62552]: ERROR nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3da47864-9e50-4957-a800-52e48d0d24b4, please check neutron logs for more information. [ 547.800556] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 547.800556] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.800556] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 547.800556] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.800556] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 547.800556] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.800556] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 547.800556] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.800556] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 547.800556] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.800556] env[62552]: ERROR nova.compute.manager raise self.value [ 547.800556] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.800556] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 547.800556] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.800556] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 547.800981] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.800981] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 547.800981] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3da47864-9e50-4957-a800-52e48d0d24b4, please check neutron logs for more information. [ 547.800981] env[62552]: ERROR nova.compute.manager [ 547.800981] env[62552]: Traceback (most recent call last): [ 547.800981] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 547.800981] env[62552]: listener.cb(fileno) [ 547.800981] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.800981] env[62552]: result = function(*args, **kwargs) [ 547.800981] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.800981] env[62552]: return func(*args, **kwargs) [ 547.800981] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.800981] env[62552]: raise e [ 547.800981] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.800981] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 547.800981] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.800981] env[62552]: created_port_ids = self._update_ports_for_instance( [ 547.800981] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.800981] env[62552]: with excutils.save_and_reraise_exception(): [ 547.800981] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.800981] env[62552]: self.force_reraise() [ 547.800981] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.800981] env[62552]: raise self.value [ 547.800981] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.800981] env[62552]: updated_port = self._update_port( [ 547.800981] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.800981] env[62552]: _ensure_no_port_binding_failure(port) [ 547.800981] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.800981] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 547.801738] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 3da47864-9e50-4957-a800-52e48d0d24b4, please check neutron logs for more information. [ 547.801738] env[62552]: Removing descriptor: 15 [ 547.802364] env[62552]: ERROR nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3da47864-9e50-4957-a800-52e48d0d24b4, please check neutron logs for more information. [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Traceback (most recent call last): [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] yield resources [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self.driver.spawn(context, instance, image_meta, [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] vm_ref = self.build_virtual_machine(instance, [ 547.802364] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] for vif in network_info: [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] return self._sync_wrapper(fn, *args, **kwargs) [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self.wait() [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self[:] = self._gt.wait() [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] return self._exit_event.wait() [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.802693] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] result = hub.switch() [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] return self.greenlet.switch() [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] result = function(*args, **kwargs) [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] return func(*args, **kwargs) [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] raise e [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] nwinfo = self.network_api.allocate_for_instance( [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] created_port_ids = self._update_ports_for_instance( [ 547.802992] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] with excutils.save_and_reraise_exception(): [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self.force_reraise() [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] raise self.value [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] updated_port = self._update_port( [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] _ensure_no_port_binding_failure(port) [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] raise exception.PortBindingFailed(port_id=port['id']) [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] nova.exception.PortBindingFailed: Binding failed for port 3da47864-9e50-4957-a800-52e48d0d24b4, please check neutron logs for more information. [ 547.803316] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] [ 547.803648] env[62552]: INFO nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Terminating instance [ 547.974649] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Acquiring lock "694aaace-af82-4eec-b959-2998d608aac1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.975021] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Lock "694aaace-af82-4eec-b959-2998d608aac1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.076186] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "1178592c-329f-4af6-94ba-07db03c07f60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.076428] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "1178592c-329f-4af6-94ba-07db03c07f60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.235208] env[62552]: DEBUG nova.compute.utils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 548.240489] env[62552]: DEBUG nova.compute.manager [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Not allocating networking since 'none' was specified. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 548.308460] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "refresh_cache-cffa534e-d51e-484c-b0e4-b27f6d094203" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.308650] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquired lock "refresh_cache-cffa534e-d51e-484c-b0e4-b27f6d094203" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.308839] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.477128] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 548.582027] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 548.741677] env[62552]: DEBUG nova.compute.manager [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 548.893990] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.917778] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f390d3ec-a717-4766-9640-b6a990b731fb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.926965] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1f9f31-0351-40ba-a2df-607f0a4ba19b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.963854] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c650bd42-31c9-4881-aec9-160bec421e14 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.970863] env[62552]: ERROR nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3175847b-1a90-4540-8a21-c434f25d077c, please check neutron logs for more information. [ 548.970863] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 548.970863] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 548.970863] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 548.970863] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.970863] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 548.970863] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.970863] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 548.970863] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.970863] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 548.970863] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.970863] env[62552]: ERROR nova.compute.manager raise self.value [ 548.970863] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.970863] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 548.970863] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.970863] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 548.971453] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.971453] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 548.971453] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3175847b-1a90-4540-8a21-c434f25d077c, please check neutron logs for more information. [ 548.971453] env[62552]: ERROR nova.compute.manager [ 548.971606] env[62552]: Traceback (most recent call last): [ 548.971904] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 548.971904] env[62552]: listener.cb(fileno) [ 548.971904] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.971904] env[62552]: result = function(*args, **kwargs) [ 548.971904] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.971904] env[62552]: return func(*args, **kwargs) [ 548.971904] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 548.971904] env[62552]: raise e [ 548.971904] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 548.971904] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 548.971904] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.971904] env[62552]: created_port_ids = self._update_ports_for_instance( [ 548.971904] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.971904] env[62552]: with excutils.save_and_reraise_exception(): [ 548.971904] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.971904] env[62552]: self.force_reraise() [ 548.971904] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.971904] env[62552]: raise self.value [ 548.971904] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.971904] env[62552]: updated_port = self._update_port( [ 548.971904] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.971904] env[62552]: _ensure_no_port_binding_failure(port) [ 548.971904] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.971904] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 548.971904] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 3175847b-1a90-4540-8a21-c434f25d077c, please check neutron logs for more information. [ 548.971904] env[62552]: Removing descriptor: 16 [ 548.972939] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5e118b-5791-4af4-a705-59fb0e3723dd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.978424] env[62552]: ERROR nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3175847b-1a90-4540-8a21-c434f25d077c, please check neutron logs for more information. [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Traceback (most recent call last): [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] yield resources [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self.driver.spawn(context, instance, image_meta, [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] vm_ref = self.build_virtual_machine(instance, [ 548.978424] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] for vif in network_info: [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] return self._sync_wrapper(fn, *args, **kwargs) [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self.wait() [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self[:] = self._gt.wait() [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] return self._exit_event.wait() [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.978808] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] result = hub.switch() [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] return self.greenlet.switch() [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] result = function(*args, **kwargs) [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] return func(*args, **kwargs) [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] raise e [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] nwinfo = self.network_api.allocate_for_instance( [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] created_port_ids = self._update_ports_for_instance( [ 548.979151] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] with excutils.save_and_reraise_exception(): [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self.force_reraise() [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] raise self.value [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] updated_port = self._update_port( [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] _ensure_no_port_binding_failure(port) [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] raise exception.PortBindingFailed(port_id=port['id']) [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] nova.exception.PortBindingFailed: Binding failed for port 3175847b-1a90-4540-8a21-c434f25d077c, please check neutron logs for more information. [ 548.979490] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] [ 548.979808] env[62552]: INFO nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Terminating instance [ 548.992441] env[62552]: DEBUG nova.compute.provider_tree [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.007540] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.053993] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.107839] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.224409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "7408a538-6091-4aa2-b2d0-a3d93840b341" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.224658] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "7408a538-6091-4aa2-b2d0-a3d93840b341" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.489908] env[62552]: DEBUG nova.compute.manager [req-bc93eaff-3f00-4de7-8f11-8a48ef897174 req-0a33a990-b749-4f44-a39f-6a3ded6047ed service nova] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Received event network-changed-3da47864-9e50-4957-a800-52e48d0d24b4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 549.490309] env[62552]: DEBUG nova.compute.manager [req-bc93eaff-3f00-4de7-8f11-8a48ef897174 req-0a33a990-b749-4f44-a39f-6a3ded6047ed service nova] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Refreshing instance network info cache due to event network-changed-3da47864-9e50-4957-a800-52e48d0d24b4. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 549.490401] env[62552]: DEBUG oslo_concurrency.lockutils [req-bc93eaff-3f00-4de7-8f11-8a48ef897174 req-0a33a990-b749-4f44-a39f-6a3ded6047ed service nova] Acquiring lock "refresh_cache-cffa534e-d51e-484c-b0e4-b27f6d094203" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.496489] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "refresh_cache-373f9654-0871-4957-abaa-d4724a9c43c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.496650] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "refresh_cache-373f9654-0871-4957-abaa-d4724a9c43c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.496814] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.498330] env[62552]: DEBUG nova.scheduler.client.report [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 549.557793] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Releasing lock "refresh_cache-cffa534e-d51e-484c-b0e4-b27f6d094203" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.558249] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 549.558515] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 549.558845] env[62552]: DEBUG oslo_concurrency.lockutils [req-bc93eaff-3f00-4de7-8f11-8a48ef897174 req-0a33a990-b749-4f44-a39f-6a3ded6047ed service nova] Acquired lock "refresh_cache-cffa534e-d51e-484c-b0e4-b27f6d094203" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.559036] env[62552]: DEBUG nova.network.neutron [req-bc93eaff-3f00-4de7-8f11-8a48ef897174 req-0a33a990-b749-4f44-a39f-6a3ded6047ed service nova] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Refreshing network info cache for port 3da47864-9e50-4957-a800-52e48d0d24b4 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 549.560227] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf8aeef4-1bae-49b2-b0f0-71638e6b6d6e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.572018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9688ac-2a49-4453-9423-f1def8ddbdf9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.597277] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cffa534e-d51e-484c-b0e4-b27f6d094203 could not be found. [ 549.597739] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 549.598051] env[62552]: INFO nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Took 0.04 seconds to destroy the instance on the hypervisor. [ 549.598516] env[62552]: DEBUG oslo.service.loopingcall [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 549.598516] env[62552]: DEBUG nova.compute.manager [-] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 549.598614] env[62552]: DEBUG nova.network.neutron [-] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 549.629340] env[62552]: DEBUG nova.network.neutron [-] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.730538] env[62552]: DEBUG nova.compute.manager [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 549.754440] env[62552]: DEBUG nova.compute.manager [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 549.794762] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 549.795013] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 549.795319] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 549.795449] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 549.795589] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 549.795728] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 549.795970] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 549.797798] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 549.798248] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 549.798591] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 549.798635] env[62552]: DEBUG nova.virt.hardware [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.799541] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5c7b44-6554-4b0a-bb7c-01bbc54d2fa1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.810715] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853d1f73-0cc1-45cb-ba75-38e548200092 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.827527] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 549.840419] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 549.844983] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cb509584-4fdc-4cb5-9c76-84fdb62a7bf7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.856417] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Created folder: OpenStack in parent group-v4. [ 549.856606] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Creating folder: Project (5e30073682004fe28ced30fb498741a4). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 549.856857] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-987aad37-8203-4b8c-8eca-a59579a6621b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.871763] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Created folder: Project (5e30073682004fe28ced30fb498741a4) in parent group-v267339. [ 549.871763] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Creating folder: Instances. Parent ref: group-v267340. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 549.872027] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44a28d28-05b3-446d-a2a9-32012f6dc61a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.880707] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Created folder: Instances in parent group-v267340. [ 549.880893] env[62552]: DEBUG oslo.service.loopingcall [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 549.881525] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 549.881525] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-01717732-a5a4-43ac-8f5c-4dc4d9ecb441 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.899055] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 549.899055] env[62552]: value = "task-1239169" [ 549.899055] env[62552]: _type = "Task" [ 549.899055] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.914047] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239169, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.005477] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.010047] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 550.018885] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.295s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.020378] env[62552]: INFO nova.compute.claims [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 550.047013] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.112340] env[62552]: DEBUG nova.network.neutron [req-bc93eaff-3f00-4de7-8f11-8a48ef897174 req-0a33a990-b749-4f44-a39f-6a3ded6047ed service nova] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.133349] env[62552]: DEBUG nova.network.neutron [-] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.284372] env[62552]: DEBUG nova.network.neutron [req-bc93eaff-3f00-4de7-8f11-8a48ef897174 req-0a33a990-b749-4f44-a39f-6a3ded6047ed service nova] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.349287] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.354120] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.410589] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239169, 'name': CreateVM_Task, 'duration_secs': 0.300682} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.410943] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 550.412087] env[62552]: DEBUG oslo_vmware.service [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c683b7-ef55-4981-bc8e-569338d78c8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.421385] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.422482] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.422482] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 550.422482] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bee72edb-afd9-409d-b4ad-a1c81edec0e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.427996] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 550.427996] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c8ef9a-9f01-d487-2e12-49b8ada8b804" [ 550.427996] env[62552]: _type = "Task" [ 550.427996] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.437432] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c8ef9a-9f01-d487-2e12-49b8ada8b804, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.532467] env[62552]: DEBUG nova.compute.utils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.533815] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 550.533987] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 550.637548] env[62552]: INFO nova.compute.manager [-] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Took 1.04 seconds to deallocate network for instance. [ 550.643116] env[62552]: DEBUG nova.compute.claims [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 550.644390] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.701639] env[62552]: DEBUG nova.policy [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5dcf36fc6c0488e817c2cb429b7dd63', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32cc16838418438482625f558462f01c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.790539] env[62552]: DEBUG oslo_concurrency.lockutils [req-bc93eaff-3f00-4de7-8f11-8a48ef897174 req-0a33a990-b749-4f44-a39f-6a3ded6047ed service nova] Releasing lock "refresh_cache-cffa534e-d51e-484c-b0e4-b27f6d094203" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.857757] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "refresh_cache-373f9654-0871-4957-abaa-d4724a9c43c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.858313] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 550.860837] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 550.861162] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8421b85-b288-4125-8ff1-8f9505abde28 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.875130] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7507613-1ad2-4af4-ab33-a3ecdd2b2380 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.908022] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 373f9654-0871-4957-abaa-d4724a9c43c9 could not be found. [ 550.908462] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 550.908800] env[62552]: INFO nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 550.909262] env[62552]: DEBUG oslo.service.loopingcall [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.909732] env[62552]: DEBUG nova.compute.manager [-] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 550.909925] env[62552]: DEBUG nova.network.neutron [-] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.941906] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.942377] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 550.942841] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.943140] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.943824] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 550.947207] env[62552]: DEBUG nova.network.neutron [-] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.947207] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fbb0b289-6fd1-496e-9941-b36d736353c1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.965546] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 550.966207] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 550.967205] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4e1bfb-537c-4408-ae05-c7277df2e908 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.977330] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d3b3e5a-26dd-43f2-b50c-ea9486ce1a4e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.982689] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 550.982689] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528b764c-478b-7561-1683-30951ec43336" [ 550.982689] env[62552]: _type = "Task" [ 550.982689] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.993084] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528b764c-478b-7561-1683-30951ec43336, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.038246] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 551.264607] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece0ce7c-2847-44c9-90b1-89f515d0fdf6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.273778] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47243ad6-98ee-4bff-82fa-a5c2227de280 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.321102] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5f79e9-178d-4b75-9275-ed27c1b34e9a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.329034] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da52a6e8-d5de-484e-8210-d60b62b9beb9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.343784] env[62552]: DEBUG nova.compute.provider_tree [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.450573] env[62552]: DEBUG nova.network.neutron [-] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.495524] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Preparing fetch location {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 551.496534] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Creating directory with path [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 551.496534] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ccd191e4-f266-4b17-bf04-ee0f6bc9142b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.500400] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Successfully created port: 184e8264-79b5-44cc-b5e1-0f56cfe69479 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.517166] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Created directory with path [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 551.518624] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Fetch image to [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 551.518624] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Downloading image file data 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee to [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk on the data store datastore2 {{(pid=62552) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 551.518624] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9772d76-495b-4bd2-962f-472fccb80d87 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.528489] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94def0b5-cafc-4d5b-b011-a22b9dafd57a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.540273] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047ac9a9-2459-4f01-bb0f-7306ca0bb21b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.582548] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5737f6-889e-4bdc-aef9-957a1e6e8ad8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.588090] env[62552]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-a889f97b-9146-4487-a239-604485e26311 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.611269] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Downloading image file data 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee to the data store datastore2 {{(pid=62552) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 551.701412] env[62552]: DEBUG oslo_vmware.rw_handles [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62552) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 551.846828] env[62552]: DEBUG nova.scheduler.client.report [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 551.955479] env[62552]: INFO nova.compute.manager [-] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Took 1.05 seconds to deallocate network for instance. [ 551.958611] env[62552]: DEBUG nova.compute.claims [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 551.958743] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.050347] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 552.098150] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.098265] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.098439] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.098629] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.098782] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.098923] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.100348] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.100436] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.100570] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.100731] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.101719] env[62552]: DEBUG nova.virt.hardware [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.101920] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce851dc-1de7-4e88-935c-c6c83c985eef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.114087] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d5c483-d75c-4931-b08b-adf563dc5625 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.358784] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.359362] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 552.363902] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.653s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.366580] env[62552]: INFO nova.compute.claims [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 552.598500] env[62552]: DEBUG oslo_vmware.rw_handles [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Completed reading data from the image iterator. {{(pid=62552) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 552.598500] env[62552]: DEBUG oslo_vmware.rw_handles [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 552.652386] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Downloaded image file data 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee to vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk on the data store datastore2 {{(pid=62552) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 552.654196] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Caching image {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 552.654693] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Copying Virtual Disk [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk to [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 552.655727] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a838f22-707e-4d19-ad38-3062115751f2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.667038] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 552.667038] env[62552]: value = "task-1239170" [ 552.667038] env[62552]: _type = "Task" [ 552.667038] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.681129] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239170, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.875497] env[62552]: DEBUG nova.compute.utils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 552.879406] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 552.879406] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 553.177575] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239170, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.184663] env[62552]: DEBUG nova.policy [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18603f735ad04cbfa2cc943ab771c7bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '77031446ea2a490ca8e350bdce5ca6c9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 553.380241] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 553.642030] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5b55a5-2535-4356-bcb8-017bbf165104 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.651246] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19e36b2-e3f1-44cd-8b0a-0adb49394382 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.694274] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31feba90-1c86-441c-bdc7-eac5960d36d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.702975] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239170, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.794742} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.705038] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Copied Virtual Disk [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk to [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 553.705231] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Deleting the datastore file [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 553.705474] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78cbe108-cfb3-4c90-991c-97088d054ba8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.707956] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.708388] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.708465] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 553.708593] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Rebuilding the list of instances to heal {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 553.714721] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20c8f87-dd4d-4535-b833-0d3914542a4f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.719715] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 553.719715] env[62552]: value = "task-1239171" [ 553.719715] env[62552]: _type = "Task" [ 553.719715] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.731536] env[62552]: DEBUG nova.compute.provider_tree [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.738520] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239171, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023475} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.738759] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 553.738959] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Moving file from [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee to [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee. {{(pid=62552) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 553.739215] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-e9e07a70-27d9-444f-b6b6-94392208061b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.745630] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 553.745630] env[62552]: value = "task-1239172" [ 553.745630] env[62552]: _type = "Task" [ 553.745630] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.754794] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239172, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.218424] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 554.218424] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 554.218424] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 554.218424] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 554.218424] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 554.218424] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 554.218710] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Didn't find any instances for network info cache update. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10357}} [ 554.218710] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.218710] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.218710] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.218838] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.219110] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.219318] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.219487] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 554.219631] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.237045] env[62552]: DEBUG nova.scheduler.client.report [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 554.258020] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239172, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.022939} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.258020] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] File moved {{(pid=62552) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 554.258020] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Cleaning up location [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8 {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 554.258020] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Deleting the datastore file [datastore2] vmware_temp/145fd963-e469-4ab5-a360-b4761ff5b5a8 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 554.258020] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5aa8602-31b0-499d-b803-d82501386729 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.266272] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 554.266272] env[62552]: value = "task-1239173" [ 554.266272] env[62552]: _type = "Task" [ 554.266272] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.275613] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239173, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.391201] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 554.429772] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 554.429946] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 554.430113] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 554.430556] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 554.430623] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 554.430741] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 554.430951] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 554.431122] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 554.431599] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 554.431795] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 554.431970] env[62552]: DEBUG nova.virt.hardware [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 554.433243] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac62b54-4fad-4f81-b624-faa1fe0172fd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.444587] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2f5e68-c180-49c4-90b3-ec6c21f4916e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.451114] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Successfully created port: c39829a3-235e-4f9e-8999-ffecbd286da1 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 554.728787] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.743987] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.743987] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 554.746432] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.400s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.747924] env[62552]: INFO nova.compute.claims [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 554.781303] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239173, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026097} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.781607] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 554.782826] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2c27ac2-dc07-463a-b1d1-27534eae04a9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.790100] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 554.790100] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52947d1f-0573-a5e4-9b65-d9e8d01c06c8" [ 554.790100] env[62552]: _type = "Task" [ 554.790100] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.798893] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52947d1f-0573-a5e4-9b65-d9e8d01c06c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.255636] env[62552]: DEBUG nova.compute.utils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 555.261965] env[62552]: ERROR nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 184e8264-79b5-44cc-b5e1-0f56cfe69479, please check neutron logs for more information. [ 555.261965] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 555.261965] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 555.261965] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 555.261965] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.261965] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 555.261965] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.261965] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 555.261965] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.261965] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 555.261965] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.261965] env[62552]: ERROR nova.compute.manager raise self.value [ 555.261965] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.261965] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 555.261965] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.261965] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 555.262560] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.262560] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 555.262560] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 184e8264-79b5-44cc-b5e1-0f56cfe69479, please check neutron logs for more information. [ 555.262560] env[62552]: ERROR nova.compute.manager [ 555.266368] env[62552]: Traceback (most recent call last): [ 555.266368] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 555.266368] env[62552]: listener.cb(fileno) [ 555.266368] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.266368] env[62552]: result = function(*args, **kwargs) [ 555.266368] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 555.266368] env[62552]: return func(*args, **kwargs) [ 555.266368] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 555.266368] env[62552]: raise e [ 555.266368] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 555.266368] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 555.266368] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.266368] env[62552]: created_port_ids = self._update_ports_for_instance( [ 555.266368] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.266368] env[62552]: with excutils.save_and_reraise_exception(): [ 555.266368] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.266368] env[62552]: self.force_reraise() [ 555.266368] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.266368] env[62552]: raise self.value [ 555.266368] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.266368] env[62552]: updated_port = self._update_port( [ 555.266368] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.266368] env[62552]: _ensure_no_port_binding_failure(port) [ 555.266368] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.266368] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 555.266368] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 184e8264-79b5-44cc-b5e1-0f56cfe69479, please check neutron logs for more information. [ 555.266368] env[62552]: Removing descriptor: 15 [ 555.267297] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 555.267297] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 555.267297] env[62552]: ERROR nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 184e8264-79b5-44cc-b5e1-0f56cfe69479, please check neutron logs for more information. [ 555.267297] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Traceback (most recent call last): [ 555.267297] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 555.267297] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] yield resources [ 555.267297] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 555.267297] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self.driver.spawn(context, instance, image_meta, [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] vm_ref = self.build_virtual_machine(instance, [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] for vif in network_info: [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] return self._sync_wrapper(fn, *args, **kwargs) [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self.wait() [ 555.267536] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self[:] = self._gt.wait() [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] return self._exit_event.wait() [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] result = hub.switch() [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] return self.greenlet.switch() [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] result = function(*args, **kwargs) [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] return func(*args, **kwargs) [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 555.267914] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] raise e [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] nwinfo = self.network_api.allocate_for_instance( [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] created_port_ids = self._update_ports_for_instance( [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] with excutils.save_and_reraise_exception(): [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self.force_reraise() [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] raise self.value [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] updated_port = self._update_port( [ 555.268301] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.268642] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] _ensure_no_port_binding_failure(port) [ 555.268642] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.268642] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] raise exception.PortBindingFailed(port_id=port['id']) [ 555.268642] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] nova.exception.PortBindingFailed: Binding failed for port 184e8264-79b5-44cc-b5e1-0f56cfe69479, please check neutron logs for more information. [ 555.268642] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] [ 555.268642] env[62552]: INFO nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Terminating instance [ 555.301243] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52947d1f-0573-a5e4-9b65-d9e8d01c06c8, 'name': SearchDatastore_Task, 'duration_secs': 0.008769} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.301642] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.302177] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6240d4e9-79ff-4c84-9d04-49879088bde4/6240d4e9-79ff-4c84-9d04-49879088bde4.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 555.302724] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e69d820-9ee2-4a87-a308-5138e71a6008 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.310863] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 555.310863] env[62552]: value = "task-1239174" [ 555.310863] env[62552]: _type = "Task" [ 555.310863] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.321737] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239174, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.391463] env[62552]: DEBUG nova.policy [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ddde0ce53d99432f9a099eff52f79d97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0910f8988bfd462a9bfaad3ffd009de6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 555.759635] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 555.777647] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Acquiring lock "refresh_cache-c56252ef-675a-4e4e-86ee-6cf36157f04d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.777923] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Acquired lock "refresh_cache-c56252ef-675a-4e4e-86ee-6cf36157f04d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.778109] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 555.785756] env[62552]: DEBUG nova.compute.manager [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Received event network-vif-deleted-3da47864-9e50-4957-a800-52e48d0d24b4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 555.786361] env[62552]: DEBUG nova.compute.manager [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Received event network-changed-3175847b-1a90-4540-8a21-c434f25d077c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 555.787862] env[62552]: DEBUG nova.compute.manager [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Refreshing instance network info cache due to event network-changed-3175847b-1a90-4540-8a21-c434f25d077c. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 555.788499] env[62552]: DEBUG oslo_concurrency.lockutils [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] Acquiring lock "refresh_cache-373f9654-0871-4957-abaa-d4724a9c43c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.788499] env[62552]: DEBUG oslo_concurrency.lockutils [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] Acquired lock "refresh_cache-373f9654-0871-4957-abaa-d4724a9c43c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.788607] env[62552]: DEBUG nova.network.neutron [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Refreshing network info cache for port 3175847b-1a90-4540-8a21-c434f25d077c {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 555.826511] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239174, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480589} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.826824] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6240d4e9-79ff-4c84-9d04-49879088bde4/6240d4e9-79ff-4c84-9d04-49879088bde4.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 555.827093] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 555.827356] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2ce2a29-94eb-4848-9d19-c210b0f2149d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.837970] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 555.837970] env[62552]: value = "task-1239175" [ 555.837970] env[62552]: _type = "Task" [ 555.837970] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.847998] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239175, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.950202] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a84348-96e7-4925-8e9c-01ef6ccfa5e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.957767] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a52342-c5b6-40f2-9d82-05c9cc915941 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.989018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec74681-3d94-437e-9ab2-8324632ebc4a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.996386] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb41261-6451-49bc-ad35-e1bcb159d5a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.010594] env[62552]: DEBUG nova.compute.provider_tree [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.343098] env[62552]: DEBUG nova.network.neutron [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.349814] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239175, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062426} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.350162] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 556.351017] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8860756-50e2-4fb7-9d4e-aa278e8856e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.372664] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 6240d4e9-79ff-4c84-9d04-49879088bde4/6240d4e9-79ff-4c84-9d04-49879088bde4.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 556.374316] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e10a0c25-1d86-4417-b64a-fdb0e6adab23 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.389744] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.390591] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Successfully created port: 6be44229-ddc9-4a45-9c5e-330e3a87577e {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 556.397198] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 556.397198] env[62552]: value = "task-1239176" [ 556.397198] env[62552]: _type = "Task" [ 556.397198] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.406370] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239176, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.477282] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.515859] env[62552]: DEBUG nova.scheduler.client.report [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 556.770586] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 556.833444] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.833749] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.834060] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.834333] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.835025] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.836484] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.838040] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.838241] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.838517] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.838754] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.838980] env[62552]: DEBUG nova.virt.hardware [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.841518] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a3785d-6ebb-4cef-ad93-7f3154855003 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.853059] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4d8b86-b917-4dbc-9eee-eff69f3f4ed2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.906435] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239176, 'name': ReconfigVM_Task, 'duration_secs': 0.273844} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.906723] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 6240d4e9-79ff-4c84-9d04-49879088bde4/6240d4e9-79ff-4c84-9d04-49879088bde4.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 556.907436] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-55d3e9d3-b8b2-4180-a5c9-874028edca3c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.913966] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 556.913966] env[62552]: value = "task-1239177" [ 556.913966] env[62552]: _type = "Task" [ 556.913966] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.922591] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239177, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.935319] env[62552]: DEBUG nova.network.neutron [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.979635] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Releasing lock "refresh_cache-c56252ef-675a-4e4e-86ee-6cf36157f04d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.980502] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 556.980502] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 556.980817] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d21a9807-5058-411c-8cc7-34192f252889 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.991061] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad8d1c4-e73c-4bdf-a104-37b023ecefc0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.014607] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c56252ef-675a-4e4e-86ee-6cf36157f04d could not be found. [ 557.014842] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 557.015031] env[62552]: INFO nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 557.015278] env[62552]: DEBUG oslo.service.loopingcall [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.015910] env[62552]: DEBUG nova.compute.manager [-] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 557.015910] env[62552]: DEBUG nova.network.neutron [-] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.019983] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.020476] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 557.025782] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.017s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.026146] env[62552]: INFO nova.compute.claims [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 557.065125] env[62552]: DEBUG nova.network.neutron [-] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.426424] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239177, 'name': Rename_Task, 'duration_secs': 0.14482} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.426751] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 557.427033] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-047079f9-92f7-4fd3-8a1f-fb0a65e983a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.434257] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 557.434257] env[62552]: value = "task-1239178" [ 557.434257] env[62552]: _type = "Task" [ 557.434257] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.443333] env[62552]: DEBUG oslo_concurrency.lockutils [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] Releasing lock "refresh_cache-373f9654-0871-4957-abaa-d4724a9c43c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.443606] env[62552]: DEBUG nova.compute.manager [req-30836e09-cb73-4e99-99b9-a08064ed43f5 req-6961d4d8-51f4-4e56-9ed8-93beef6c56e2 service nova] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Received event network-vif-deleted-3175847b-1a90-4540-8a21-c434f25d077c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 557.444081] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239178, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.537723] env[62552]: DEBUG nova.compute.utils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 557.547722] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 557.547722] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 557.568376] env[62552]: DEBUG nova.network.neutron [-] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.747757] env[62552]: DEBUG nova.policy [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae637e1f0eb9473caeed3d9657ce85cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae984e014d9649a5a05770ed69a68f15', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 557.947073] env[62552]: DEBUG oslo_vmware.api [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239178, 'name': PowerOnVM_Task, 'duration_secs': 0.480659} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.947403] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 557.947633] env[62552]: INFO nova.compute.manager [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Took 8.19 seconds to spawn the instance on the hypervisor. [ 557.947792] env[62552]: DEBUG nova.compute.manager [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 557.948691] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ff1e99-b1f6-4e22-83d7-b301fc5f1314 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.046707] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 558.072282] env[62552]: INFO nova.compute.manager [-] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Took 1.06 seconds to deallocate network for instance. [ 558.081719] env[62552]: DEBUG nova.compute.claims [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 558.082060] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.252092] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7931fde8-1c46-4e9f-8b40-87fbc4f0da75 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.261922] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f0f3fa-7cf3-4f38-b5b7-6e05ee76c77b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.302267] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c317a4-e01a-4fd5-ac80-7a48e997bbe0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.310524] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c017a211-8e79-46cc-86bc-b073f4ff65e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.325551] env[62552]: DEBUG nova.compute.provider_tree [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.474564] env[62552]: INFO nova.compute.manager [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Took 14.49 seconds to build instance. [ 558.606080] env[62552]: ERROR nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c39829a3-235e-4f9e-8999-ffecbd286da1, please check neutron logs for more information. [ 558.606080] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 558.606080] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 558.606080] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 558.606080] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.606080] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 558.606080] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.606080] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 558.606080] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.606080] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 558.606080] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.606080] env[62552]: ERROR nova.compute.manager raise self.value [ 558.606080] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.606080] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 558.606080] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.606080] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 558.606830] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.606830] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 558.606830] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c39829a3-235e-4f9e-8999-ffecbd286da1, please check neutron logs for more information. [ 558.606830] env[62552]: ERROR nova.compute.manager [ 558.606830] env[62552]: Traceback (most recent call last): [ 558.606830] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 558.606830] env[62552]: listener.cb(fileno) [ 558.606830] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.606830] env[62552]: result = function(*args, **kwargs) [ 558.606830] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 558.606830] env[62552]: return func(*args, **kwargs) [ 558.606830] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 558.606830] env[62552]: raise e [ 558.606830] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 558.606830] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 558.606830] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.606830] env[62552]: created_port_ids = self._update_ports_for_instance( [ 558.606830] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.606830] env[62552]: with excutils.save_and_reraise_exception(): [ 558.606830] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.606830] env[62552]: self.force_reraise() [ 558.606830] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.606830] env[62552]: raise self.value [ 558.606830] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.606830] env[62552]: updated_port = self._update_port( [ 558.606830] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.606830] env[62552]: _ensure_no_port_binding_failure(port) [ 558.606830] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.606830] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 558.607536] env[62552]: nova.exception.PortBindingFailed: Binding failed for port c39829a3-235e-4f9e-8999-ffecbd286da1, please check neutron logs for more information. [ 558.607536] env[62552]: Removing descriptor: 16 [ 558.607536] env[62552]: ERROR nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c39829a3-235e-4f9e-8999-ffecbd286da1, please check neutron logs for more information. [ 558.607536] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Traceback (most recent call last): [ 558.607536] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 558.607536] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] yield resources [ 558.607536] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 558.607536] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self.driver.spawn(context, instance, image_meta, [ 558.607536] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 558.607536] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.607536] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.607536] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] vm_ref = self.build_virtual_machine(instance, [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] for vif in network_info: [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] return self._sync_wrapper(fn, *args, **kwargs) [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self.wait() [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self[:] = self._gt.wait() [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] return self._exit_event.wait() [ 558.608048] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] result = hub.switch() [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] return self.greenlet.switch() [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] result = function(*args, **kwargs) [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] return func(*args, **kwargs) [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] raise e [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] nwinfo = self.network_api.allocate_for_instance( [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.608539] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] created_port_ids = self._update_ports_for_instance( [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] with excutils.save_and_reraise_exception(): [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self.force_reraise() [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] raise self.value [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] updated_port = self._update_port( [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] _ensure_no_port_binding_failure(port) [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.608867] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] raise exception.PortBindingFailed(port_id=port['id']) [ 558.609185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] nova.exception.PortBindingFailed: Binding failed for port c39829a3-235e-4f9e-8999-ffecbd286da1, please check neutron logs for more information. [ 558.609185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] [ 558.609185] env[62552]: INFO nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Terminating instance [ 558.828768] env[62552]: DEBUG nova.scheduler.client.report [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 558.848156] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Successfully created port: 41d2b6d1-7f0e-4300-a691-4d0486c2dd35 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.978636] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa2b0fdf-10e8-4f9e-b519-ac73df6a7deb tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "6240d4e9-79ff-4c84-9d04-49879088bde4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.003s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.057374] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 559.088692] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 559.088692] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 559.088692] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 559.089007] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 559.089007] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 559.089007] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 559.089007] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 559.089007] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 559.091383] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 559.091529] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 559.091701] env[62552]: DEBUG nova.virt.hardware [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 559.092645] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596cf0f9-8a6d-46ec-a67b-ad7dde565cc6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.101890] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab075b4-0a25-4197-a3a6-2ebfb03256df {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.118458] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "refresh_cache-2f52f72d-176a-4278-9592-b06ee4807df1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.118637] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquired lock "refresh_cache-2f52f72d-176a-4278-9592-b06ee4807df1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.118821] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 559.334581] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.335107] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 559.341048] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.230s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.341048] env[62552]: INFO nova.compute.claims [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 559.626025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Acquiring lock "fab97d93-2322-469d-84fa-812e1d491030" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.626025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Lock "fab97d93-2322-469d-84fa-812e1d491030" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.680477] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.845251] env[62552]: DEBUG nova.compute.utils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 559.847996] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.850212] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 559.850587] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 559.963217] env[62552]: DEBUG nova.policy [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5375a838f8b46e78d4720003aad7cff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '27ca01f1d32c4a8ca929bb223f5a3c5e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 560.036403] env[62552]: ERROR nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6be44229-ddc9-4a45-9c5e-330e3a87577e, please check neutron logs for more information. [ 560.036403] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 560.036403] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 560.036403] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 560.036403] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.036403] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 560.036403] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.036403] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 560.036403] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.036403] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 560.036403] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.036403] env[62552]: ERROR nova.compute.manager raise self.value [ 560.036403] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.036403] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 560.036403] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.036403] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 560.037296] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.037296] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 560.037296] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6be44229-ddc9-4a45-9c5e-330e3a87577e, please check neutron logs for more information. [ 560.037296] env[62552]: ERROR nova.compute.manager [ 560.038470] env[62552]: Traceback (most recent call last): [ 560.038677] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 560.038677] env[62552]: listener.cb(fileno) [ 560.038677] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.038677] env[62552]: result = function(*args, **kwargs) [ 560.038677] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.038677] env[62552]: return func(*args, **kwargs) [ 560.038677] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 560.038677] env[62552]: raise e [ 560.039803] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 560.039803] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 560.039803] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.039803] env[62552]: created_port_ids = self._update_ports_for_instance( [ 560.039803] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.039803] env[62552]: with excutils.save_and_reraise_exception(): [ 560.039803] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.039803] env[62552]: self.force_reraise() [ 560.039803] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.039803] env[62552]: raise self.value [ 560.039803] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.039803] env[62552]: updated_port = self._update_port( [ 560.039803] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.039803] env[62552]: _ensure_no_port_binding_failure(port) [ 560.039803] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.039803] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 560.039803] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 6be44229-ddc9-4a45-9c5e-330e3a87577e, please check neutron logs for more information. [ 560.039803] env[62552]: Removing descriptor: 15 [ 560.039803] env[62552]: ERROR nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6be44229-ddc9-4a45-9c5e-330e3a87577e, please check neutron logs for more information. [ 560.039803] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Traceback (most recent call last): [ 560.039803] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] yield resources [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self.driver.spawn(context, instance, image_meta, [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] vm_ref = self.build_virtual_machine(instance, [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] for vif in network_info: [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.040502] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] return self._sync_wrapper(fn, *args, **kwargs) [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self.wait() [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self[:] = self._gt.wait() [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] return self._exit_event.wait() [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] result = hub.switch() [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] return self.greenlet.switch() [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] result = function(*args, **kwargs) [ 560.041824] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] return func(*args, **kwargs) [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] raise e [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] nwinfo = self.network_api.allocate_for_instance( [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] created_port_ids = self._update_ports_for_instance( [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] with excutils.save_and_reraise_exception(): [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self.force_reraise() [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.042574] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] raise self.value [ 560.047405] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.047405] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] updated_port = self._update_port( [ 560.047405] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.047405] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] _ensure_no_port_binding_failure(port) [ 560.047405] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.047405] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] raise exception.PortBindingFailed(port_id=port['id']) [ 560.047405] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] nova.exception.PortBindingFailed: Binding failed for port 6be44229-ddc9-4a45-9c5e-330e3a87577e, please check neutron logs for more information. [ 560.047405] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] [ 560.047405] env[62552]: INFO nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Terminating instance [ 560.127295] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 560.351198] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 560.361412] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Releasing lock "refresh_cache-2f52f72d-176a-4278-9592-b06ee4807df1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.362403] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 560.362938] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 560.363280] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50f619fa-79df-4fb1-a0d4-25d6df4b9db1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.373115] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead58d9d-289e-4f4b-ad6d-0c39886c7a2f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.401967] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2f52f72d-176a-4278-9592-b06ee4807df1 could not be found. [ 560.402237] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 560.402419] env[62552]: INFO nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 560.402659] env[62552]: DEBUG oslo.service.loopingcall [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 560.405822] env[62552]: DEBUG nova.compute.manager [-] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 560.406492] env[62552]: DEBUG nova.network.neutron [-] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 560.453044] env[62552]: DEBUG nova.network.neutron [-] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.545202] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Acquiring lock "refresh_cache-05687d2d-0760-4c5d-a231-4ef7c2d440e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.545387] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Acquired lock "refresh_cache-05687d2d-0760-4c5d-a231-4ef7c2d440e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.545557] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.562128] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa72c1d-1cf4-4cd2-9995-69726db7f95a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.572696] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d03ee4-af4c-4352-bd7a-c08c42a8e194 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.606688] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f39045a-cbe4-4f30-8137-ec9b7d53e49a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.615798] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4aad3d-cdbe-4fe6-82f2-c0dcafa6a795 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.629118] env[62552]: DEBUG nova.compute.provider_tree [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.637249] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Successfully created port: dea477e3-cfd3-4ba6-8a50-0199829c3f07 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 560.651423] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.771982] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "c4a725a7-a14c-44ec-a81a-3e965d7e0817" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.772274] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "c4a725a7-a14c-44ec-a81a-3e965d7e0817" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.933250] env[62552]: DEBUG nova.compute.manager [req-681cb13a-dd29-4438-a7b5-133a1aa6647a req-658fd518-f666-4740-a329-003d7723fba9 service nova] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Received event network-changed-184e8264-79b5-44cc-b5e1-0f56cfe69479 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 560.933736] env[62552]: DEBUG nova.compute.manager [req-681cb13a-dd29-4438-a7b5-133a1aa6647a req-658fd518-f666-4740-a329-003d7723fba9 service nova] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Refreshing instance network info cache due to event network-changed-184e8264-79b5-44cc-b5e1-0f56cfe69479. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 560.933994] env[62552]: DEBUG oslo_concurrency.lockutils [req-681cb13a-dd29-4438-a7b5-133a1aa6647a req-658fd518-f666-4740-a329-003d7723fba9 service nova] Acquiring lock "refresh_cache-c56252ef-675a-4e4e-86ee-6cf36157f04d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.934767] env[62552]: DEBUG oslo_concurrency.lockutils [req-681cb13a-dd29-4438-a7b5-133a1aa6647a req-658fd518-f666-4740-a329-003d7723fba9 service nova] Acquired lock "refresh_cache-c56252ef-675a-4e4e-86ee-6cf36157f04d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.934953] env[62552]: DEBUG nova.network.neutron [req-681cb13a-dd29-4438-a7b5-133a1aa6647a req-658fd518-f666-4740-a329-003d7723fba9 service nova] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Refreshing network info cache for port 184e8264-79b5-44cc-b5e1-0f56cfe69479 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 560.957250] env[62552]: DEBUG nova.network.neutron [-] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.080591] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.132868] env[62552]: DEBUG nova.scheduler.client.report [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 561.156857] env[62552]: DEBUG nova.compute.manager [None req-ade9ab6b-142c-401a-ab8f-22c59c48d465 tempest-ServerDiagnosticsV248Test-1931939617 tempest-ServerDiagnosticsV248Test-1931939617-project-admin] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 561.159337] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4630a6c-3cc9-4fab-b69c-b8dd7bfc12c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.169013] env[62552]: INFO nova.compute.manager [None req-ade9ab6b-142c-401a-ab8f-22c59c48d465 tempest-ServerDiagnosticsV248Test-1931939617 tempest-ServerDiagnosticsV248Test-1931939617-project-admin] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Retrieving diagnostics [ 561.169013] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed3022da-0a0e-4e5d-b4a5-d6dccf77b83c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.327973] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.363931] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 561.394856] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 561.395122] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 561.395281] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 561.395459] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 561.395604] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 561.395746] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 561.396318] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 561.396521] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 561.396696] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 561.396859] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 561.397049] env[62552]: DEBUG nova.virt.hardware [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 561.398802] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24be2c7f-7b11-4539-b643-bc7ab76d775e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.407488] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ad5b22-d0d8-455a-b52e-0443672cfa07 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.459072] env[62552]: INFO nova.compute.manager [-] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Took 1.05 seconds to deallocate network for instance. [ 561.462231] env[62552]: DEBUG nova.compute.claims [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 561.462428] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.472602] env[62552]: DEBUG nova.network.neutron [req-681cb13a-dd29-4438-a7b5-133a1aa6647a req-658fd518-f666-4740-a329-003d7723fba9 service nova] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.611391] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "a44de380-895c-4f06-835e-595917a25be1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.611995] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "a44de380-895c-4f06-835e-595917a25be1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.619939] env[62552]: DEBUG nova.network.neutron [req-681cb13a-dd29-4438-a7b5-133a1aa6647a req-658fd518-f666-4740-a329-003d7723fba9 service nova] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.635191] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "27e4d3de-3e44-4293-9b04-e511c9f02b42" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.635506] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "27e4d3de-3e44-4293-9b04-e511c9f02b42" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.641731] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.642198] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 561.648141] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.297s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.648141] env[62552]: INFO nova.compute.claims [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.650984] env[62552]: ERROR nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35, please check neutron logs for more information. [ 561.650984] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 561.650984] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.650984] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 561.650984] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.650984] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 561.650984] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.650984] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 561.650984] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.650984] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 561.650984] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.650984] env[62552]: ERROR nova.compute.manager raise self.value [ 561.650984] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.650984] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 561.650984] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.650984] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 561.651515] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.651515] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 561.651515] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35, please check neutron logs for more information. [ 561.651515] env[62552]: ERROR nova.compute.manager [ 561.651515] env[62552]: Traceback (most recent call last): [ 561.651515] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 561.651515] env[62552]: listener.cb(fileno) [ 561.651515] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.651515] env[62552]: result = function(*args, **kwargs) [ 561.651515] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.651515] env[62552]: return func(*args, **kwargs) [ 561.651515] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 561.651515] env[62552]: raise e [ 561.651515] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.651515] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 561.651515] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.651515] env[62552]: created_port_ids = self._update_ports_for_instance( [ 561.651515] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.651515] env[62552]: with excutils.save_and_reraise_exception(): [ 561.651515] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.651515] env[62552]: self.force_reraise() [ 561.651515] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.651515] env[62552]: raise self.value [ 561.651515] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.651515] env[62552]: updated_port = self._update_port( [ 561.651515] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.651515] env[62552]: _ensure_no_port_binding_failure(port) [ 561.651515] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.651515] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 561.652370] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35, please check neutron logs for more information. [ 561.652370] env[62552]: Removing descriptor: 19 [ 561.652370] env[62552]: ERROR nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35, please check neutron logs for more information. [ 561.652370] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Traceback (most recent call last): [ 561.652370] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 561.652370] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] yield resources [ 561.652370] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 561.652370] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self.driver.spawn(context, instance, image_meta, [ 561.652370] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 561.652370] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.652370] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.652370] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] vm_ref = self.build_virtual_machine(instance, [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] for vif in network_info: [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] return self._sync_wrapper(fn, *args, **kwargs) [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self.wait() [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self[:] = self._gt.wait() [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] return self._exit_event.wait() [ 561.652722] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] result = hub.switch() [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] return self.greenlet.switch() [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] result = function(*args, **kwargs) [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] return func(*args, **kwargs) [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] raise e [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] nwinfo = self.network_api.allocate_for_instance( [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.653945] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] created_port_ids = self._update_ports_for_instance( [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] with excutils.save_and_reraise_exception(): [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self.force_reraise() [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] raise self.value [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] updated_port = self._update_port( [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] _ensure_no_port_binding_failure(port) [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.654392] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] raise exception.PortBindingFailed(port_id=port['id']) [ 561.654737] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] nova.exception.PortBindingFailed: Binding failed for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35, please check neutron logs for more information. [ 561.654737] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] [ 561.654737] env[62552]: INFO nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Terminating instance [ 561.703439] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "92b9d5a7-e5ed-480e-b68e-b589e6d00832" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.703746] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "92b9d5a7-e5ed-480e-b68e-b589e6d00832" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.830390] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Releasing lock "refresh_cache-05687d2d-0760-4c5d-a231-4ef7c2d440e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.830817] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 561.831021] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 561.831314] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ef16d2c-e8c4-4d63-bf96-7a9bd5186d9d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.840709] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cf7405-9bb8-4bcd-965b-a026fd5e85cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.863474] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 05687d2d-0760-4c5d-a231-4ef7c2d440e6 could not be found. [ 561.864047] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 561.864047] env[62552]: INFO nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 561.864191] env[62552]: DEBUG oslo.service.loopingcall [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 561.864430] env[62552]: DEBUG nova.compute.manager [-] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 561.864522] env[62552]: DEBUG nova.network.neutron [-] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 561.891736] env[62552]: DEBUG nova.network.neutron [-] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.123885] env[62552]: DEBUG oslo_concurrency.lockutils [req-681cb13a-dd29-4438-a7b5-133a1aa6647a req-658fd518-f666-4740-a329-003d7723fba9 service nova] Releasing lock "refresh_cache-c56252ef-675a-4e4e-86ee-6cf36157f04d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.124703] env[62552]: DEBUG nova.compute.manager [req-681cb13a-dd29-4438-a7b5-133a1aa6647a req-658fd518-f666-4740-a329-003d7723fba9 service nova] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Received event network-vif-deleted-184e8264-79b5-44cc-b5e1-0f56cfe69479 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 562.153938] env[62552]: DEBUG nova.compute.utils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 562.155548] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 562.155724] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 562.158143] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Acquiring lock "refresh_cache-b3dcf7cf-5807-4483-8ca0-a245931c7dc3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.158230] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Acquired lock "refresh_cache-b3dcf7cf-5807-4483-8ca0-a245931c7dc3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.158390] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.294119] env[62552]: DEBUG nova.policy [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c732a4a15f9b4b16b241c09f4ba373e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3daa5253a0cc4538903b7cabc2f98fa2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 562.393724] env[62552]: DEBUG nova.network.neutron [-] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.553973] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Acquiring lock "a34ea7fb-bd17-4353-806d-4cd2f5ee169d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.554233] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Lock "a34ea7fb-bd17-4353-806d-4cd2f5ee169d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.662043] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 562.710548] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.835552] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.897260] env[62552]: INFO nova.compute.manager [-] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Took 1.03 seconds to deallocate network for instance. [ 562.901570] env[62552]: DEBUG nova.compute.claims [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 562.903339] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.940821] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa5af11-c271-48f4-a354-b16367a453b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.949327] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd61af2-95c5-4460-a9b0-34f92dfb0133 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.980743] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Successfully created port: b4978a45-38c6-4be0-97b4-b5d4937c68fe {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 562.983213] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e70bf4-1246-43c9-bed3-ff36794c2f23 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.992646] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be754552-bc98-4d5c-963a-f93483c2e194 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.007284] env[62552]: DEBUG nova.compute.provider_tree [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.227083] env[62552]: ERROR nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dea477e3-cfd3-4ba6-8a50-0199829c3f07, please check neutron logs for more information. [ 563.227083] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.227083] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 563.227083] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.227083] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.227083] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.227083] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.227083] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.227083] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.227083] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 563.227083] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.227083] env[62552]: ERROR nova.compute.manager raise self.value [ 563.227083] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.227083] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.227083] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.227083] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.227753] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.227753] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.227753] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dea477e3-cfd3-4ba6-8a50-0199829c3f07, please check neutron logs for more information. [ 563.227753] env[62552]: ERROR nova.compute.manager [ 563.227753] env[62552]: Traceback (most recent call last): [ 563.227753] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.227753] env[62552]: listener.cb(fileno) [ 563.227753] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.227753] env[62552]: result = function(*args, **kwargs) [ 563.227753] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.227753] env[62552]: return func(*args, **kwargs) [ 563.227753] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 563.227753] env[62552]: raise e [ 563.227753] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 563.227753] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 563.227753] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.227753] env[62552]: created_port_ids = self._update_ports_for_instance( [ 563.227753] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.227753] env[62552]: with excutils.save_and_reraise_exception(): [ 563.227753] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.227753] env[62552]: self.force_reraise() [ 563.227753] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.227753] env[62552]: raise self.value [ 563.227753] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.227753] env[62552]: updated_port = self._update_port( [ 563.227753] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.227753] env[62552]: _ensure_no_port_binding_failure(port) [ 563.227753] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.227753] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.228739] env[62552]: nova.exception.PortBindingFailed: Binding failed for port dea477e3-cfd3-4ba6-8a50-0199829c3f07, please check neutron logs for more information. [ 563.228739] env[62552]: Removing descriptor: 16 [ 563.228739] env[62552]: ERROR nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dea477e3-cfd3-4ba6-8a50-0199829c3f07, please check neutron logs for more information. [ 563.228739] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] Traceback (most recent call last): [ 563.228739] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 563.228739] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] yield resources [ 563.228739] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 563.228739] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self.driver.spawn(context, instance, image_meta, [ 563.228739] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 563.228739] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.228739] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.228739] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] vm_ref = self.build_virtual_machine(instance, [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] for vif in network_info: [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] return self._sync_wrapper(fn, *args, **kwargs) [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self.wait() [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self[:] = self._gt.wait() [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] return self._exit_event.wait() [ 563.229120] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] result = hub.switch() [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] return self.greenlet.switch() [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] result = function(*args, **kwargs) [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] return func(*args, **kwargs) [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] raise e [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] nwinfo = self.network_api.allocate_for_instance( [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.229481] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] created_port_ids = self._update_ports_for_instance( [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] with excutils.save_and_reraise_exception(): [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self.force_reraise() [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] raise self.value [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] updated_port = self._update_port( [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] _ensure_no_port_binding_failure(port) [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.229843] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] raise exception.PortBindingFailed(port_id=port['id']) [ 563.230207] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] nova.exception.PortBindingFailed: Binding failed for port dea477e3-cfd3-4ba6-8a50-0199829c3f07, please check neutron logs for more information. [ 563.230207] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] [ 563.230207] env[62552]: INFO nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Terminating instance [ 563.338550] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Releasing lock "refresh_cache-b3dcf7cf-5807-4483-8ca0-a245931c7dc3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.338971] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 563.339206] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 563.339502] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a8cd747-26d4-4532-8aab-d135bf7c637b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.348157] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a6f3d3-f8ce-44dd-b629-f78306f33937 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.371662] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b3dcf7cf-5807-4483-8ca0-a245931c7dc3 could not be found. [ 563.371885] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 563.372077] env[62552]: INFO nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 563.372313] env[62552]: DEBUG oslo.service.loopingcall [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.372506] env[62552]: DEBUG nova.compute.manager [-] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 563.372597] env[62552]: DEBUG nova.network.neutron [-] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.414943] env[62552]: DEBUG nova.network.neutron [-] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.512670] env[62552]: DEBUG nova.scheduler.client.report [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 563.670931] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 563.708311] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 563.708601] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 563.708762] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 563.708945] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 563.710141] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 563.710141] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 563.710141] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 563.710358] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 563.710394] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 563.713127] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 563.713127] env[62552]: DEBUG nova.virt.hardware [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 563.713127] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87efd546-cd7c-4c07-aa8e-18156cbf8b0b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.720677] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87f705b-d720-477a-a865-d961fa8d7e1a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.734134] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Acquiring lock "refresh_cache-694aaace-af82-4eec-b959-2998d608aac1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.734347] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Acquired lock "refresh_cache-694aaace-af82-4eec-b959-2998d608aac1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.734530] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.918633] env[62552]: DEBUG nova.network.neutron [-] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.018019] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.018019] env[62552]: DEBUG nova.compute.manager [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 564.018988] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.375s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.164130] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "58c91f56-a683-4a62-bd45-80c627622621" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.165422] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "58c91f56-a683-4a62-bd45-80c627622621" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.264496] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.274427] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Received event network-changed-c39829a3-235e-4f9e-8999-ffecbd286da1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 564.274595] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Refreshing instance network info cache due to event network-changed-c39829a3-235e-4f9e-8999-ffecbd286da1. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 564.275799] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Acquiring lock "refresh_cache-2f52f72d-176a-4278-9592-b06ee4807df1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.275799] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Acquired lock "refresh_cache-2f52f72d-176a-4278-9592-b06ee4807df1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.275799] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Refreshing network info cache for port c39829a3-235e-4f9e-8999-ffecbd286da1 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 564.420767] env[62552]: INFO nova.compute.manager [-] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Took 1.05 seconds to deallocate network for instance. [ 564.424796] env[62552]: DEBUG nova.compute.claims [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 564.424971] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.463115] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.524569] env[62552]: DEBUG nova.compute.utils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 564.531394] env[62552]: DEBUG nova.compute.manager [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Not allocating networking since 'none' was specified. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 564.722093] env[62552]: ERROR nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b4978a45-38c6-4be0-97b4-b5d4937c68fe, please check neutron logs for more information. [ 564.722093] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 564.722093] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 564.722093] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 564.722093] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.722093] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 564.722093] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.722093] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 564.722093] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.722093] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 564.722093] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.722093] env[62552]: ERROR nova.compute.manager raise self.value [ 564.722093] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.722093] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 564.722093] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.722093] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 564.722559] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.722559] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 564.722559] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b4978a45-38c6-4be0-97b4-b5d4937c68fe, please check neutron logs for more information. [ 564.722559] env[62552]: ERROR nova.compute.manager [ 564.722559] env[62552]: Traceback (most recent call last): [ 564.722559] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 564.722559] env[62552]: listener.cb(fileno) [ 564.722559] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.722559] env[62552]: result = function(*args, **kwargs) [ 564.722559] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.722559] env[62552]: return func(*args, **kwargs) [ 564.722559] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 564.722559] env[62552]: raise e [ 564.722559] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 564.722559] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 564.722559] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.722559] env[62552]: created_port_ids = self._update_ports_for_instance( [ 564.722559] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.722559] env[62552]: with excutils.save_and_reraise_exception(): [ 564.722559] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.722559] env[62552]: self.force_reraise() [ 564.722559] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.722559] env[62552]: raise self.value [ 564.722559] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.722559] env[62552]: updated_port = self._update_port( [ 564.722559] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.722559] env[62552]: _ensure_no_port_binding_failure(port) [ 564.722559] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.722559] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 564.723455] env[62552]: nova.exception.PortBindingFailed: Binding failed for port b4978a45-38c6-4be0-97b4-b5d4937c68fe, please check neutron logs for more information. [ 564.723455] env[62552]: Removing descriptor: 19 [ 564.723594] env[62552]: ERROR nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b4978a45-38c6-4be0-97b4-b5d4937c68fe, please check neutron logs for more information. [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Traceback (most recent call last): [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] yield resources [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self.driver.spawn(context, instance, image_meta, [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] vm_ref = self.build_virtual_machine(instance, [ 564.723594] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] for vif in network_info: [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] return self._sync_wrapper(fn, *args, **kwargs) [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self.wait() [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self[:] = self._gt.wait() [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] return self._exit_event.wait() [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.724041] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] result = hub.switch() [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] return self.greenlet.switch() [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] result = function(*args, **kwargs) [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] return func(*args, **kwargs) [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] raise e [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] nwinfo = self.network_api.allocate_for_instance( [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] created_port_ids = self._update_ports_for_instance( [ 564.724540] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] with excutils.save_and_reraise_exception(): [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self.force_reraise() [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] raise self.value [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] updated_port = self._update_port( [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] _ensure_no_port_binding_failure(port) [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] raise exception.PortBindingFailed(port_id=port['id']) [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] nova.exception.PortBindingFailed: Binding failed for port b4978a45-38c6-4be0-97b4-b5d4937c68fe, please check neutron logs for more information. [ 564.724949] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] [ 564.725331] env[62552]: INFO nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Terminating instance [ 564.796931] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17120e75-0691-45cc-ac88-f52a510f70fb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.806674] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b95983-62fa-431b-a8d8-fae68929d63b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.810587] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.839540] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6455bd1-e4aa-4f55-9405-aaab64bb212f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.849165] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb9f8fb-c235-4696-ba72-b9dca8c48b64 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.863211] env[62552]: DEBUG nova.compute.provider_tree [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.939654] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.968612] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Releasing lock "refresh_cache-694aaace-af82-4eec-b959-2998d608aac1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.969024] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 564.969213] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.969480] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7cfe3bfa-46b2-42e6-846b-1dd2a15bbbc6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.977870] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f45808-e029-40da-8092-da65623543cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.000133] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 694aaace-af82-4eec-b959-2998d608aac1 could not be found. [ 565.000389] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 565.000568] env[62552]: INFO nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 565.001061] env[62552]: DEBUG oslo.service.loopingcall [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.001061] env[62552]: DEBUG nova.compute.manager [-] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 565.001330] env[62552]: DEBUG nova.network.neutron [-] [instance: 694aaace-af82-4eec-b959-2998d608aac1] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 565.027387] env[62552]: DEBUG nova.network.neutron [-] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.032861] env[62552]: DEBUG nova.compute.manager [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 565.233804] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-1178592c-329f-4af6-94ba-07db03c07f60" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.233992] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-1178592c-329f-4af6-94ba-07db03c07f60" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.234181] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.269889] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "9952d8f6-ee78-4c2d-b147-8c08c027f440" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.270186] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "9952d8f6-ee78-4c2d-b147-8c08c027f440" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.366991] env[62552]: DEBUG nova.scheduler.client.report [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 565.442699] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Releasing lock "refresh_cache-2f52f72d-176a-4278-9592-b06ee4807df1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.442985] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Received event network-vif-deleted-c39829a3-235e-4f9e-8999-ffecbd286da1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 565.443205] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Received event network-changed-6be44229-ddc9-4a45-9c5e-330e3a87577e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 565.443362] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Refreshing instance network info cache due to event network-changed-6be44229-ddc9-4a45-9c5e-330e3a87577e. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 565.443568] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Acquiring lock "refresh_cache-05687d2d-0760-4c5d-a231-4ef7c2d440e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.443704] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Acquired lock "refresh_cache-05687d2d-0760-4c5d-a231-4ef7c2d440e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.443855] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Refreshing network info cache for port 6be44229-ddc9-4a45-9c5e-330e3a87577e {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 565.530550] env[62552]: DEBUG nova.network.neutron [-] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.761814] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.880036] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.861s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.880389] env[62552]: ERROR nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3da47864-9e50-4957-a800-52e48d0d24b4, please check neutron logs for more information. [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Traceback (most recent call last): [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self.driver.spawn(context, instance, image_meta, [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] vm_ref = self.build_virtual_machine(instance, [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.880389] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] for vif in network_info: [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] return self._sync_wrapper(fn, *args, **kwargs) [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self.wait() [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self[:] = self._gt.wait() [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] return self._exit_event.wait() [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] result = hub.switch() [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.880735] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] return self.greenlet.switch() [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] result = function(*args, **kwargs) [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] return func(*args, **kwargs) [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] raise e [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] nwinfo = self.network_api.allocate_for_instance( [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] created_port_ids = self._update_ports_for_instance( [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] with excutils.save_and_reraise_exception(): [ 565.881113] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] self.force_reraise() [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] raise self.value [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] updated_port = self._update_port( [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] _ensure_no_port_binding_failure(port) [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] raise exception.PortBindingFailed(port_id=port['id']) [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] nova.exception.PortBindingFailed: Binding failed for port 3da47864-9e50-4957-a800-52e48d0d24b4, please check neutron logs for more information. [ 565.881455] env[62552]: ERROR nova.compute.manager [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] [ 565.881721] env[62552]: DEBUG nova.compute.utils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Binding failed for port 3da47864-9e50-4957-a800-52e48d0d24b4, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.882595] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.924s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.892967] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Build of instance cffa534e-d51e-484c-b0e4-b27f6d094203 was re-scheduled: Binding failed for port 3da47864-9e50-4957-a800-52e48d0d24b4, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 565.893712] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 565.894062] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "refresh_cache-cffa534e-d51e-484c-b0e4-b27f6d094203" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.894296] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquired lock "refresh_cache-cffa534e-d51e-484c-b0e4-b27f6d094203" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.896254] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.966909] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.034471] env[62552]: INFO nova.compute.manager [-] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Took 1.03 seconds to deallocate network for instance. [ 566.038273] env[62552]: DEBUG nova.compute.claims [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 566.038774] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.040721] env[62552]: DEBUG nova.compute.manager [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 566.082895] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.083179] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.083336] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.083515] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.083806] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.083967] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.084189] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.084450] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.084655] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.084821] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.084990] env[62552]: DEBUG nova.virt.hardware [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.085910] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da42032-c785-431e-93b1-17bd6d50befb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.096509] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3135fa25-37e3-41a9-8348-cf95f037143d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.112206] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 566.118289] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Creating folder: Project (2c202d08c9b04309ab69f9c91fbb555c). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 566.119288] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.121363] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4178c453-5a1c-4151-b336-c36340a53f98 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.131516] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Created folder: Project (2c202d08c9b04309ab69f9c91fbb555c) in parent group-v267339. [ 566.131708] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Creating folder: Instances. Parent ref: group-v267343. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 566.132326] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd4aad81-d055-4dfd-8e85-12c0ee07d3fd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.141071] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Created folder: Instances in parent group-v267343. [ 566.141338] env[62552]: DEBUG oslo.service.loopingcall [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.141534] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 566.141734] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08ede6de-ea55-46c1-9213-eef605379f47 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.160359] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 566.160359] env[62552]: value = "task-1239181" [ 566.160359] env[62552]: _type = "Task" [ 566.160359] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.170081] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239181, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.312881] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.427845] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.470915] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-1178592c-329f-4af6-94ba-07db03c07f60" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.471779] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 566.472526] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 566.478022] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f84b4aa9-2f24-4894-a9fd-5a038f63663e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.492939] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d745d16-a4ef-4f8f-af0b-402051c305db {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.516336] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1178592c-329f-4af6-94ba-07db03c07f60 could not be found. [ 566.516636] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 566.516858] env[62552]: INFO nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Took 0.04 seconds to destroy the instance on the hypervisor. [ 566.517141] env[62552]: DEBUG oslo.service.loopingcall [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.520088] env[62552]: DEBUG nova.compute.manager [-] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 566.520226] env[62552]: DEBUG nova.network.neutron [-] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.559601] env[62552]: DEBUG nova.network.neutron [-] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.617970] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.670718] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239181, 'name': CreateVM_Task, 'duration_secs': 0.325182} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.673210] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 566.673850] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.674071] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.674684] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 566.674872] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eedb2c08-cd78-4131-b60c-60fa14e70981 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.679835] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 566.679835] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526f79b0-5769-964f-69ef-d75e517453f8" [ 566.679835] env[62552]: _type = "Task" [ 566.679835] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.688164] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526f79b0-5769-964f-69ef-d75e517453f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.720560] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5be9ef5-7b4c-4b3b-9dc0-315b13d03456 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.729313] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed69330-a1ab-4bb4-8713-8cb0aca192b5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.764446] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1238378d-c47d-4a10-8d55-fe3b5ff21f22 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.774783] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a560e6a-9599-40df-84f9-0907738d7cd3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.790764] env[62552]: DEBUG nova.compute.provider_tree [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.815439] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Releasing lock "refresh_cache-05687d2d-0760-4c5d-a231-4ef7c2d440e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.815439] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Received event network-vif-deleted-6be44229-ddc9-4a45-9c5e-330e3a87577e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 566.815714] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Received event network-changed-41d2b6d1-7f0e-4300-a691-4d0486c2dd35 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 566.815714] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Refreshing instance network info cache due to event network-changed-41d2b6d1-7f0e-4300-a691-4d0486c2dd35. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 566.815986] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Acquiring lock "refresh_cache-b3dcf7cf-5807-4483-8ca0-a245931c7dc3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.816231] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Acquired lock "refresh_cache-b3dcf7cf-5807-4483-8ca0-a245931c7dc3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.816383] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Refreshing network info cache for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 567.061942] env[62552]: DEBUG nova.network.neutron [-] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.124159] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Releasing lock "refresh_cache-cffa534e-d51e-484c-b0e4-b27f6d094203" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.124410] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 567.124575] env[62552]: DEBUG nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 567.124740] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 567.146301] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.198472] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526f79b0-5769-964f-69ef-d75e517453f8, 'name': SearchDatastore_Task, 'duration_secs': 0.008962} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.198718] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.199235] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 567.199235] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.199359] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.199512] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 567.199773] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95c69bf1-9d43-4276-a712-08cdbe86018c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.207879] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 567.208087] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 567.209023] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b30100a8-9a49-458f-8b82-0f7175a6b0de {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.214152] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 567.214152] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a906e6-99b5-ccec-4991-e95b501b987e" [ 567.214152] env[62552]: _type = "Task" [ 567.214152] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.222668] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a906e6-99b5-ccec-4991-e95b501b987e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.229435] env[62552]: DEBUG nova.compute.manager [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Received event network-vif-deleted-dea477e3-cfd3-4ba6-8a50-0199829c3f07 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 567.229641] env[62552]: DEBUG nova.compute.manager [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Received event network-changed-b4978a45-38c6-4be0-97b4-b5d4937c68fe {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 567.229804] env[62552]: DEBUG nova.compute.manager [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Refreshing instance network info cache due to event network-changed-b4978a45-38c6-4be0-97b4-b5d4937c68fe. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 567.230010] env[62552]: DEBUG oslo_concurrency.lockutils [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] Acquiring lock "refresh_cache-1178592c-329f-4af6-94ba-07db03c07f60" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.230159] env[62552]: DEBUG oslo_concurrency.lockutils [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] Acquired lock "refresh_cache-1178592c-329f-4af6-94ba-07db03c07f60" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.230316] env[62552]: DEBUG nova.network.neutron [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Refreshing network info cache for port b4978a45-38c6-4be0-97b4-b5d4937c68fe {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 567.296903] env[62552]: DEBUG nova.scheduler.client.report [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 567.341102] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.425232] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.567603] env[62552]: INFO nova.compute.manager [-] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Took 1.05 seconds to deallocate network for instance. [ 567.569707] env[62552]: DEBUG nova.compute.claims [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 567.569707] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.648856] env[62552]: DEBUG nova.network.neutron [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.727281] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a906e6-99b5-ccec-4991-e95b501b987e, 'name': SearchDatastore_Task, 'duration_secs': 0.008437} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.728116] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1465d238-325e-48f0-a2c3-e6174ebbe019 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.733882] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 567.733882] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c5e402-b1ec-2ec2-ce47-55e2e8be7dde" [ 567.733882] env[62552]: _type = "Task" [ 567.733882] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.743541] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c5e402-b1ec-2ec2-ce47-55e2e8be7dde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.767575] env[62552]: DEBUG nova.network.neutron [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.802549] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.803722] env[62552]: ERROR nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3175847b-1a90-4540-8a21-c434f25d077c, please check neutron logs for more information. [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Traceback (most recent call last): [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self.driver.spawn(context, instance, image_meta, [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] vm_ref = self.build_virtual_machine(instance, [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.803722] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] for vif in network_info: [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] return self._sync_wrapper(fn, *args, **kwargs) [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self.wait() [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self[:] = self._gt.wait() [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] return self._exit_event.wait() [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] result = hub.switch() [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.804195] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] return self.greenlet.switch() [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] result = function(*args, **kwargs) [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] return func(*args, **kwargs) [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] raise e [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] nwinfo = self.network_api.allocate_for_instance( [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] created_port_ids = self._update_ports_for_instance( [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] with excutils.save_and_reraise_exception(): [ 567.804670] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] self.force_reraise() [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] raise self.value [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] updated_port = self._update_port( [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] _ensure_no_port_binding_failure(port) [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] raise exception.PortBindingFailed(port_id=port['id']) [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] nova.exception.PortBindingFailed: Binding failed for port 3175847b-1a90-4540-8a21-c434f25d077c, please check neutron logs for more information. [ 567.805157] env[62552]: ERROR nova.compute.manager [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] [ 567.805936] env[62552]: DEBUG nova.compute.utils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Binding failed for port 3175847b-1a90-4540-8a21-c434f25d077c, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 567.806744] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Build of instance 373f9654-0871-4957-abaa-d4724a9c43c9 was re-scheduled: Binding failed for port 3175847b-1a90-4540-8a21-c434f25d077c, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 567.807182] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 567.807418] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "refresh_cache-373f9654-0871-4957-abaa-d4724a9c43c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.807562] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "refresh_cache-373f9654-0871-4957-abaa-d4724a9c43c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.807718] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 567.812483] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.085s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.812657] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.812805] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 567.813915] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.731s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.818032] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a519d31-681b-4a2b-90dc-bac091d0e67e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.826815] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2192bc88-9e6b-4394-baa1-404e6b12811e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.842227] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09006f4-8d0e-4883-97d2-74bdede3d17e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.848712] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a761389a-4c60-4426-bf01-a575b258779b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.887105] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181544MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 567.887105] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.928241] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Releasing lock "refresh_cache-b3dcf7cf-5807-4483-8ca0-a245931c7dc3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.928529] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Received event network-vif-deleted-41d2b6d1-7f0e-4300-a691-4d0486c2dd35 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 567.928714] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Received event network-changed-dea477e3-cfd3-4ba6-8a50-0199829c3f07 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 567.928873] env[62552]: DEBUG nova.compute.manager [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Refreshing instance network info cache due to event network-changed-dea477e3-cfd3-4ba6-8a50-0199829c3f07. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 567.929138] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Acquiring lock "refresh_cache-694aaace-af82-4eec-b959-2998d608aac1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.929283] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Acquired lock "refresh_cache-694aaace-af82-4eec-b959-2998d608aac1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.930277] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Refreshing network info cache for port dea477e3-cfd3-4ba6-8a50-0199829c3f07 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 567.974182] env[62552]: DEBUG nova.network.neutron [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.153447] env[62552]: INFO nova.compute.manager [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: cffa534e-d51e-484c-b0e4-b27f6d094203] Took 1.03 seconds to deallocate network for instance. [ 568.248641] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c5e402-b1ec-2ec2-ce47-55e2e8be7dde, 'name': SearchDatastore_Task, 'duration_secs': 0.008449} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.249109] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.249475] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 7408a538-6091-4aa2-b2d0-a3d93840b341/7408a538-6091-4aa2-b2d0-a3d93840b341.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 568.250009] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a23a11f-4758-4150-994e-af966f4e6db9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.259022] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 568.259022] env[62552]: value = "task-1239182" [ 568.259022] env[62552]: _type = "Task" [ 568.259022] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.273023] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239182, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.353601] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.480731] env[62552]: DEBUG oslo_concurrency.lockutils [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] Releasing lock "refresh_cache-1178592c-329f-4af6-94ba-07db03c07f60" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.481050] env[62552]: DEBUG nova.compute.manager [req-11e28e51-005b-4127-bc96-67f5eaad9e3c req-b447a1ac-ea11-424a-8d7b-aeefb6d7117a service nova] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Received event network-vif-deleted-b4978a45-38c6-4be0-97b4-b5d4937c68fe {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 568.563168] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.704830] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dffb4ab-e788-4747-940d-3fc8356af248 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.717549] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5d0ada-bf5f-41f1-a6bd-b2ede6e7f942 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.769505] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad78f4a4-e8e3-42c0-9566-b3da7466731a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.778548] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239182, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488332} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.780796] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 7408a538-6091-4aa2-b2d0-a3d93840b341/7408a538-6091-4aa2-b2d0-a3d93840b341.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 568.781050] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 568.781466] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cf3145e-5d65-48e3-90f1-dc30f3975b76 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.784159] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e4375b-7bf9-4371-84c2-4accbdbd95a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.799769] env[62552]: DEBUG nova.compute.provider_tree [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.802527] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 568.802527] env[62552]: value = "task-1239183" [ 568.802527] env[62552]: _type = "Task" [ 568.802527] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.810858] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239183, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.853107] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.890040] env[62552]: DEBUG nova.network.neutron [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.205588] env[62552]: INFO nova.scheduler.client.report [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Deleted allocations for instance cffa534e-d51e-484c-b0e4-b27f6d094203 [ 569.304528] env[62552]: DEBUG nova.scheduler.client.report [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 569.323606] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239183, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056004} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.324563] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 569.326443] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2489fe-430f-406c-b877-4d08e7e69cdf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.355321] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] 7408a538-6091-4aa2-b2d0-a3d93840b341/7408a538-6091-4aa2-b2d0-a3d93840b341.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 569.356255] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "refresh_cache-373f9654-0871-4957-abaa-d4724a9c43c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.356497] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 569.356703] env[62552]: DEBUG nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 569.356952] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 569.358780] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c380e6db-0a9b-491b-be96-760893b90e28 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.381059] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 569.381059] env[62552]: value = "task-1239184" [ 569.381059] env[62552]: _type = "Task" [ 569.381059] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.393700] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239184, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.397218] env[62552]: DEBUG oslo_concurrency.lockutils [req-34f5ca5e-efd9-4026-89dc-e254b7beaaf1 req-84b9cf2b-90d8-41b3-95bb-25cbb64beca3 service nova] Releasing lock "refresh_cache-694aaace-af82-4eec-b959-2998d608aac1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.424819] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.717451] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b6f0839f-fb0e-4ae9-b6d3-b8c38c5d45b5 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "cffa534e-d51e-484c-b0e4-b27f6d094203" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.700s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.811611] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.998s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.812579] env[62552]: ERROR nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 184e8264-79b5-44cc-b5e1-0f56cfe69479, please check neutron logs for more information. [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Traceback (most recent call last): [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self.driver.spawn(context, instance, image_meta, [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] vm_ref = self.build_virtual_machine(instance, [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.812579] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] for vif in network_info: [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] return self._sync_wrapper(fn, *args, **kwargs) [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self.wait() [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self[:] = self._gt.wait() [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] return self._exit_event.wait() [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] result = hub.switch() [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.812934] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] return self.greenlet.switch() [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] result = function(*args, **kwargs) [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] return func(*args, **kwargs) [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] raise e [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] nwinfo = self.network_api.allocate_for_instance( [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] created_port_ids = self._update_ports_for_instance( [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] with excutils.save_and_reraise_exception(): [ 569.813305] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] self.force_reraise() [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] raise self.value [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] updated_port = self._update_port( [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] _ensure_no_port_binding_failure(port) [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] raise exception.PortBindingFailed(port_id=port['id']) [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] nova.exception.PortBindingFailed: Binding failed for port 184e8264-79b5-44cc-b5e1-0f56cfe69479, please check neutron logs for more information. [ 569.813617] env[62552]: ERROR nova.compute.manager [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] [ 569.816976] env[62552]: DEBUG nova.compute.utils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Binding failed for port 184e8264-79b5-44cc-b5e1-0f56cfe69479, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 569.816976] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.163s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.816976] env[62552]: INFO nova.compute.claims [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 569.824859] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Build of instance c56252ef-675a-4e4e-86ee-6cf36157f04d was re-scheduled: Binding failed for port 184e8264-79b5-44cc-b5e1-0f56cfe69479, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 569.824859] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 569.824859] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Acquiring lock "refresh_cache-c56252ef-675a-4e4e-86ee-6cf36157f04d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.824859] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Acquired lock "refresh_cache-c56252ef-675a-4e4e-86ee-6cf36157f04d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.825702] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 569.893337] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239184, 'name': ReconfigVM_Task, 'duration_secs': 0.264604} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.894912] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Reconfigured VM instance instance-0000000a to attach disk [datastore2] 7408a538-6091-4aa2-b2d0-a3d93840b341/7408a538-6091-4aa2-b2d0-a3d93840b341.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 569.894912] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d63a821-2277-448f-95ac-91bcf883ea26 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.902023] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 569.902023] env[62552]: value = "task-1239185" [ 569.902023] env[62552]: _type = "Task" [ 569.902023] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.913579] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239185, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.927456] env[62552]: DEBUG nova.network.neutron [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.223494] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 570.393391] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.415633] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239185, 'name': Rename_Task, 'duration_secs': 0.139465} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.417091] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 570.417091] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe0ba393-c5dc-45a0-9688-982cb4d15c32 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.423574] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 570.423574] env[62552]: value = "task-1239186" [ 570.423574] env[62552]: _type = "Task" [ 570.423574] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.433318] env[62552]: INFO nova.compute.manager [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 373f9654-0871-4957-abaa-d4724a9c43c9] Took 1.08 seconds to deallocate network for instance. [ 570.439390] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239186, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.645687] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.768516] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.951677] env[62552]: DEBUG oslo_vmware.api [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239186, 'name': PowerOnVM_Task, 'duration_secs': 0.498688} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.955127] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 570.955205] env[62552]: INFO nova.compute.manager [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Took 4.91 seconds to spawn the instance on the hypervisor. [ 570.955485] env[62552]: DEBUG nova.compute.manager [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 570.956800] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a75794-7bda-438d-b96e-be357ee825af {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.150514] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Releasing lock "refresh_cache-c56252ef-675a-4e4e-86ee-6cf36157f04d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.151163] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 571.151163] env[62552]: DEBUG nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 571.151163] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 571.157165] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36b16ef-b9cc-47b8-824a-b90e947ba3a0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.164768] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ff9dbc-82f1-40b2-b9ac-8705abce47a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.211736] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.213480] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186dbdcf-0935-45d7-a0f2-6cdb06970354 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.223762] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a73ddd-ae62-4cd3-88ac-de14688b518f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.242281] env[62552]: DEBUG nova.compute.provider_tree [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.478909] env[62552]: INFO nova.scheduler.client.report [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleted allocations for instance 373f9654-0871-4957-abaa-d4724a9c43c9 [ 571.490782] env[62552]: INFO nova.compute.manager [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Took 21.26 seconds to build instance. [ 571.717120] env[62552]: DEBUG nova.network.neutron [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.745543] env[62552]: DEBUG nova.scheduler.client.report [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 571.991714] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0282a70-363e-40d7-8229-e45a4eb0cbf9 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "373f9654-0871-4957-abaa-d4724a9c43c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.504s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.994849] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c8a12a3c-aeee-4fab-8489-6945ce81c4d7 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "7408a538-6091-4aa2-b2d0-a3d93840b341" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.770s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.222842] env[62552]: INFO nova.compute.manager [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] [instance: c56252ef-675a-4e4e-86ee-6cf36157f04d] Took 1.07 seconds to deallocate network for instance. [ 572.255681] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.258387] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 572.259970] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.797s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.495028] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 572.500250] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 572.600857] env[62552]: DEBUG nova.compute.manager [None req-745a8b15-7402-40ce-a860-cd18bb4d06b4 tempest-ServerDiagnosticsV248Test-1931939617 tempest-ServerDiagnosticsV248Test-1931939617-project-admin] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 572.604210] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ba38d8-bc80-400a-bcdb-7ec02fc68dd4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.613348] env[62552]: INFO nova.compute.manager [None req-745a8b15-7402-40ce-a860-cd18bb4d06b4 tempest-ServerDiagnosticsV248Test-1931939617 tempest-ServerDiagnosticsV248Test-1931939617-project-admin] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Retrieving diagnostics [ 572.614212] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc5d51f-493a-49aa-b7e8-914036872b09 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.769708] env[62552]: DEBUG nova.compute.utils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.769708] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 572.769708] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 573.040110] env[62552]: DEBUG nova.policy [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ce06290c30845ef9fa1e5630e8b14cd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d63975891474a4aa2d80dab97aec84e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 573.043808] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.046488] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.126222] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02ab846-5c76-4a2d-afa4-f137796a3c1d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.136667] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cbf4b0-cea9-4be1-b261-7df7f59b92dc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.169095] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440a5a42-1211-498d-9bbb-a19f6ac7ddd0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.177086] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deae37f6-3722-47db-9e33-69a7c9fc711c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.192036] env[62552]: DEBUG nova.compute.provider_tree [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.260468] env[62552]: INFO nova.scheduler.client.report [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Deleted allocations for instance c56252ef-675a-4e4e-86ee-6cf36157f04d [ 573.270116] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 573.462985] env[62552]: INFO nova.compute.manager [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Rebuilding instance [ 573.518493] env[62552]: DEBUG nova.compute.manager [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 573.519433] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b386368a-6386-42b3-9586-0cad320de0f2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.695094] env[62552]: DEBUG nova.scheduler.client.report [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 573.779251] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1cd212f6-4c3f-4136-be1d-939e5a808873 tempest-ServerDiagnosticsNegativeTest-1872501533 tempest-ServerDiagnosticsNegativeTest-1872501533-project-member] Lock "c56252ef-675a-4e4e-86ee-6cf36157f04d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.722s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.201949] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.942s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.202185] env[62552]: ERROR nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c39829a3-235e-4f9e-8999-ffecbd286da1, please check neutron logs for more information. [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Traceback (most recent call last): [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self.driver.spawn(context, instance, image_meta, [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] vm_ref = self.build_virtual_machine(instance, [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.202185] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] for vif in network_info: [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] return self._sync_wrapper(fn, *args, **kwargs) [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self.wait() [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self[:] = self._gt.wait() [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] return self._exit_event.wait() [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] result = hub.switch() [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.202507] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] return self.greenlet.switch() [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] result = function(*args, **kwargs) [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] return func(*args, **kwargs) [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] raise e [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] nwinfo = self.network_api.allocate_for_instance( [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] created_port_ids = self._update_ports_for_instance( [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] with excutils.save_and_reraise_exception(): [ 574.202852] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] self.force_reraise() [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] raise self.value [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] updated_port = self._update_port( [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] _ensure_no_port_binding_failure(port) [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] raise exception.PortBindingFailed(port_id=port['id']) [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] nova.exception.PortBindingFailed: Binding failed for port c39829a3-235e-4f9e-8999-ffecbd286da1, please check neutron logs for more information. [ 574.203190] env[62552]: ERROR nova.compute.manager [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] [ 574.203480] env[62552]: DEBUG nova.compute.utils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Binding failed for port c39829a3-235e-4f9e-8999-ffecbd286da1, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 574.205299] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.303s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.209513] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Build of instance 2f52f72d-176a-4278-9592-b06ee4807df1 was re-scheduled: Binding failed for port c39829a3-235e-4f9e-8999-ffecbd286da1, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 574.210086] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 574.210319] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "refresh_cache-2f52f72d-176a-4278-9592-b06ee4807df1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.210460] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquired lock "refresh_cache-2f52f72d-176a-4278-9592-b06ee4807df1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.210613] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.286829] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 574.290709] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 574.327091] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.330275] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.330485] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.330742] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.330930] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.331097] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.331709] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.331709] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.331709] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.331899] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.332388] env[62552]: DEBUG nova.virt.hardware [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.335694] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02f42b0-ddfa-4460-92db-20572fb806d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.344551] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62bf01ef-a87d-4741-998a-5eadd8b534b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.541136] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 574.541136] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efe090c0-6b8e-4650-b7bf-71ca3e6d7596 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.549382] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 574.549382] env[62552]: value = "task-1239187" [ 574.549382] env[62552]: _type = "Task" [ 574.549382] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.558205] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239187, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.661220] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Successfully created port: 6e71a8c2-6b86-47c7-85cf-a31825c39bc4 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.776112] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.824506] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.986687] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eba7629-33e2-4830-82ba-d4b16b16065d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.995133] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41fdaaec-a5e2-4790-8535-9313cdce55c2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.035113] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad3265f-5c5a-4d67-bfd5-75d47af99ec9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.045430] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Acquiring lock "729eba7d-5f14-4506-8fee-24f6eb685b12" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.045574] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Lock "729eba7d-5f14-4506-8fee-24f6eb685b12" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.046881] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48490e26-2d1a-4fc3-ad19-bfb85cce2dda {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.068906] env[62552]: DEBUG nova.compute.provider_tree [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.073225] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239187, 'name': PowerOffVM_Task, 'duration_secs': 0.123314} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.073727] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 575.073946] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 575.075282] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e11ebc-311d-4bbb-8a76-f34b9eb0f9cd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.081695] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 575.081922] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b521a84-a3b9-40fd-b05a-6ea8e39f024c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.107065] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 575.107300] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 575.107478] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Deleting the datastore file [datastore2] 7408a538-6091-4aa2-b2d0-a3d93840b341 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 575.107739] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c515d085-4fae-43f3-a943-f4bb0d537898 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.114420] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 575.114420] env[62552]: value = "task-1239189" [ 575.114420] env[62552]: _type = "Task" [ 575.114420] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.123308] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239189, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.180614] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.578025] env[62552]: DEBUG nova.scheduler.client.report [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 575.631830] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239189, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103994} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.631830] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 575.631830] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 575.632060] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 575.683382] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Releasing lock "refresh_cache-2f52f72d-176a-4278-9592-b06ee4807df1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.683758] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 575.686914] env[62552]: DEBUG nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 575.686914] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 575.732084] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.764096] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Acquiring lock "e22c0fef-3d0d-4339-be0b-f129b3628b80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.764096] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Lock "e22c0fef-3d0d-4339-be0b-f129b3628b80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.084490] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.879s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.085175] env[62552]: ERROR nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6be44229-ddc9-4a45-9c5e-330e3a87577e, please check neutron logs for more information. [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Traceback (most recent call last): [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self.driver.spawn(context, instance, image_meta, [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] vm_ref = self.build_virtual_machine(instance, [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.085175] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] for vif in network_info: [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] return self._sync_wrapper(fn, *args, **kwargs) [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self.wait() [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self[:] = self._gt.wait() [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] return self._exit_event.wait() [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] result = hub.switch() [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.085516] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] return self.greenlet.switch() [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] result = function(*args, **kwargs) [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] return func(*args, **kwargs) [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] raise e [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] nwinfo = self.network_api.allocate_for_instance( [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] created_port_ids = self._update_ports_for_instance( [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] with excutils.save_and_reraise_exception(): [ 576.085830] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] self.force_reraise() [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] raise self.value [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] updated_port = self._update_port( [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] _ensure_no_port_binding_failure(port) [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] raise exception.PortBindingFailed(port_id=port['id']) [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] nova.exception.PortBindingFailed: Binding failed for port 6be44229-ddc9-4a45-9c5e-330e3a87577e, please check neutron logs for more information. [ 576.086180] env[62552]: ERROR nova.compute.manager [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] [ 576.086448] env[62552]: DEBUG nova.compute.utils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Binding failed for port 6be44229-ddc9-4a45-9c5e-330e3a87577e, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 576.087910] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.663s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.091815] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Build of instance 05687d2d-0760-4c5d-a231-4ef7c2d440e6 was re-scheduled: Binding failed for port 6be44229-ddc9-4a45-9c5e-330e3a87577e, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 576.092264] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 576.092495] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Acquiring lock "refresh_cache-05687d2d-0760-4c5d-a231-4ef7c2d440e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.092662] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Acquired lock "refresh_cache-05687d2d-0760-4c5d-a231-4ef7c2d440e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.093706] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.235037] env[62552]: DEBUG nova.network.neutron [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.568684] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquiring lock "6240d4e9-79ff-4c84-9d04-49879088bde4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.568916] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "6240d4e9-79ff-4c84-9d04-49879088bde4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.569135] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquiring lock "6240d4e9-79ff-4c84-9d04-49879088bde4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.569322] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "6240d4e9-79ff-4c84-9d04-49879088bde4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.569499] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "6240d4e9-79ff-4c84-9d04-49879088bde4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.575037] env[62552]: INFO nova.compute.manager [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Terminating instance [ 576.580690] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Acquiring lock "2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.580690] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Lock "2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.656360] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.703250] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 576.703451] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 576.703565] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.703738] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 576.703876] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.705489] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 576.705489] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 576.705489] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 576.705489] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 576.705489] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 576.705804] env[62552]: DEBUG nova.virt.hardware [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 576.706445] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e93346c-6d38-4675-9945-451865f2780a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.725889] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86996045-5dfe-4371-9cb8-27a000a536e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.751546] env[62552]: INFO nova.compute.manager [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 2f52f72d-176a-4278-9592-b06ee4807df1] Took 1.07 seconds to deallocate network for instance. [ 576.753511] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 576.759408] env[62552]: DEBUG oslo.service.loopingcall [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.762434] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 576.763081] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8dbf9670-a425-4bbc-9c57-3ae9c3f2d355 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.786449] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 576.786449] env[62552]: value = "task-1239190" [ 576.786449] env[62552]: _type = "Task" [ 576.786449] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.800654] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239190, 'name': CreateVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.016022] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce52d625-a01f-4678-ba6e-559dbf461e91 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.024745] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec796cfc-c568-4bbf-b7cb-92e5e533e039 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.068503] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bbc0ab-1b6d-48b2-809a-a580598ef820 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.075734] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1959206-31bd-4f03-a83d-05359898b224 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.092024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquiring lock "refresh_cache-6240d4e9-79ff-4c84-9d04-49879088bde4" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.092024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquired lock "refresh_cache-6240d4e9-79ff-4c84-9d04-49879088bde4" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.092024] env[62552]: DEBUG nova.network.neutron [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 577.092500] env[62552]: DEBUG nova.compute.provider_tree [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.234277] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.303918] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239190, 'name': CreateVM_Task, 'duration_secs': 0.282006} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.304732] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 577.306972] env[62552]: DEBUG oslo_vmware.service [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c76aaa7-1d35-4128-b9a6-d60a34f7665f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.313393] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.313720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.314214] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 577.314615] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2eed69fd-c6a2-442c-aa9a-47fd29bd4310 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.320816] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 577.320816] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1d417-3d05-a454-de31-9586bbc73bb4" [ 577.320816] env[62552]: _type = "Task" [ 577.320816] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.333119] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1d417-3d05-a454-de31-9586bbc73bb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.600940] env[62552]: DEBUG nova.scheduler.client.report [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 577.662201] env[62552]: DEBUG nova.network.neutron [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.737863] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Releasing lock "refresh_cache-05687d2d-0760-4c5d-a231-4ef7c2d440e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.738145] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 577.738330] env[62552]: DEBUG nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 577.739227] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 577.795885] env[62552]: INFO nova.scheduler.client.report [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Deleted allocations for instance 2f52f72d-176a-4278-9592-b06ee4807df1 [ 577.804716] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.837503] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.838029] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 577.838628] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.838628] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.838802] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 577.839336] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ccdbacd3-3b28-45cc-97ff-6642e2c07161 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.849920] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 577.850156] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 577.851477] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c71024-90f9-4777-8699-0b102eb8cddb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.859384] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf4072a1-ed1e-4565-b0fa-9fa82539d269 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.865239] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 577.865239] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c256b5-1dda-d5c1-8a74-356d373785a6" [ 577.865239] env[62552]: _type = "Task" [ 577.865239] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.874631] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c256b5-1dda-d5c1-8a74-356d373785a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.890827] env[62552]: DEBUG nova.network.neutron [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.109202] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.020s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.110979] env[62552]: ERROR nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35, please check neutron logs for more information. [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Traceback (most recent call last): [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self.driver.spawn(context, instance, image_meta, [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] vm_ref = self.build_virtual_machine(instance, [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.110979] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] for vif in network_info: [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] return self._sync_wrapper(fn, *args, **kwargs) [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self.wait() [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self[:] = self._gt.wait() [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] return self._exit_event.wait() [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] result = hub.switch() [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.111368] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] return self.greenlet.switch() [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] result = function(*args, **kwargs) [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] return func(*args, **kwargs) [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] raise e [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] nwinfo = self.network_api.allocate_for_instance( [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] created_port_ids = self._update_ports_for_instance( [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] with excutils.save_and_reraise_exception(): [ 578.113110] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] self.force_reraise() [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] raise self.value [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] updated_port = self._update_port( [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] _ensure_no_port_binding_failure(port) [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] raise exception.PortBindingFailed(port_id=port['id']) [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] nova.exception.PortBindingFailed: Binding failed for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35, please check neutron logs for more information. [ 578.113566] env[62552]: ERROR nova.compute.manager [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] [ 578.114166] env[62552]: DEBUG nova.compute.utils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Binding failed for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 578.116813] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.077s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.122711] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Build of instance b3dcf7cf-5807-4483-8ca0-a245931c7dc3 was re-scheduled: Binding failed for port 41d2b6d1-7f0e-4300-a691-4d0486c2dd35, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 578.123972] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 578.126582] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Acquiring lock "refresh_cache-b3dcf7cf-5807-4483-8ca0-a245931c7dc3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.126582] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Acquired lock "refresh_cache-b3dcf7cf-5807-4483-8ca0-a245931c7dc3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.126582] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.305031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bd0a0198-814f-438f-86ef-e7e9f75b97b1 tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "2f52f72d-176a-4278-9592-b06ee4807df1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.619s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.306897] env[62552]: DEBUG nova.network.neutron [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.377270] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Preparing fetch location {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 578.377510] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Creating directory with path [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 578.377756] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5a4af21-510a-406b-b9a1-4e564b958271 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.394225] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Releasing lock "refresh_cache-6240d4e9-79ff-4c84-9d04-49879088bde4" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.394653] env[62552]: DEBUG nova.compute.manager [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 578.394838] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 578.395738] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed8723c-8dd4-48f0-be65-6552105be454 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.405156] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 578.407094] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f1ddfb1-76a3-42d6-9172-76a77314ba31 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.409869] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Created directory with path [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 578.410237] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Fetch image to [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 578.410237] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Downloading image file data 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee to [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk on the data store datastore1 {{(pid=62552) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 578.411301] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87bbbb8-5ce3-431a-9144-96ee7aaf51f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.419678] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c65074-c4f0-47b2-967d-d38e8002b135 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.424109] env[62552]: DEBUG oslo_vmware.api [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 578.424109] env[62552]: value = "task-1239191" [ 578.424109] env[62552]: _type = "Task" [ 578.424109] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.435012] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229c1bb5-1357-465e-82d6-270ece1f0867 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.443393] env[62552]: DEBUG oslo_vmware.api [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239191, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.480203] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63daa9a-af23-4959-8817-ba1248b3dfd7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.490673] env[62552]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-7b527672-727c-4463-94de-2999676ac3ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.535452] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Downloading image file data 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee to the data store datastore1 {{(pid=62552) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 578.564023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Acquiring lock "b6bff07b-f316-4f26-9623-d50328df1ac0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.564023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Lock "b6bff07b-f316-4f26-9623-d50328df1ac0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.616538] env[62552]: DEBUG oslo_vmware.rw_handles [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62552) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 578.810160] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 578.814208] env[62552]: INFO nova.compute.manager [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] [instance: 05687d2d-0760-4c5d-a231-4ef7c2d440e6] Took 1.08 seconds to deallocate network for instance. [ 578.920259] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.934512] env[62552]: DEBUG oslo_vmware.api [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239191, 'name': PowerOffVM_Task, 'duration_secs': 0.190844} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.935341] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 578.935341] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 578.935341] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90b21da9-3ef6-4d8f-a59d-6927f38d169a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.973747] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 578.973957] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 578.974156] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Deleting the datastore file [datastore2] 6240d4e9-79ff-4c84-9d04-49879088bde4 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 578.976682] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34c4445f-9e09-48c7-9ce1-4b9c3a2a0509 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.983251] env[62552]: DEBUG oslo_vmware.api [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for the task: (returnval){ [ 578.983251] env[62552]: value = "task-1239193" [ 578.983251] env[62552]: _type = "Task" [ 578.983251] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.995788] env[62552]: DEBUG oslo_vmware.api [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239193, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.015206] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ca8658-6720-4583-a750-2c316d2e50aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.021239] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defa18c0-9cdf-43da-908e-fa03f94cebc6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.056875] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d8e5d5-3744-4475-ace7-fa2137d92859 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.066390] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ea3663-3fda-4e8f-b011-ed665a53afd0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.080207] env[62552]: DEBUG nova.compute.provider_tree [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.136728] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.349536] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.370492] env[62552]: ERROR nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4, please check neutron logs for more information. [ 579.370492] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.370492] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 579.370492] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.370492] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.370492] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.370492] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.370492] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.370492] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.370492] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 579.370492] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.370492] env[62552]: ERROR nova.compute.manager raise self.value [ 579.370492] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.370492] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.370492] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.370492] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.371021] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.371021] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.371021] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4, please check neutron logs for more information. [ 579.371021] env[62552]: ERROR nova.compute.manager [ 579.371021] env[62552]: Traceback (most recent call last): [ 579.371021] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.371021] env[62552]: listener.cb(fileno) [ 579.371021] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.371021] env[62552]: result = function(*args, **kwargs) [ 579.371021] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.371021] env[62552]: return func(*args, **kwargs) [ 579.371021] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 579.371021] env[62552]: raise e [ 579.371021] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 579.371021] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 579.371021] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.371021] env[62552]: created_port_ids = self._update_ports_for_instance( [ 579.371021] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.371021] env[62552]: with excutils.save_and_reraise_exception(): [ 579.371021] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.371021] env[62552]: self.force_reraise() [ 579.371021] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.371021] env[62552]: raise self.value [ 579.371021] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.371021] env[62552]: updated_port = self._update_port( [ 579.371021] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.371021] env[62552]: _ensure_no_port_binding_failure(port) [ 579.371021] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.371021] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.371789] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4, please check neutron logs for more information. [ 579.371789] env[62552]: Removing descriptor: 19 [ 579.371789] env[62552]: ERROR nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4, please check neutron logs for more information. [ 579.371789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] Traceback (most recent call last): [ 579.371789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 579.371789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] yield resources [ 579.371789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 579.371789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self.driver.spawn(context, instance, image_meta, [ 579.371789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 579.371789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.371789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.371789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] vm_ref = self.build_virtual_machine(instance, [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] for vif in network_info: [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] return self._sync_wrapper(fn, *args, **kwargs) [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self.wait() [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self[:] = self._gt.wait() [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] return self._exit_event.wait() [ 579.372218] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] result = hub.switch() [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] return self.greenlet.switch() [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] result = function(*args, **kwargs) [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] return func(*args, **kwargs) [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] raise e [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] nwinfo = self.network_api.allocate_for_instance( [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.372584] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] created_port_ids = self._update_ports_for_instance( [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] with excutils.save_and_reraise_exception(): [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self.force_reraise() [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] raise self.value [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] updated_port = self._update_port( [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] _ensure_no_port_binding_failure(port) [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.373789] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] raise exception.PortBindingFailed(port_id=port['id']) [ 579.374169] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] nova.exception.PortBindingFailed: Binding failed for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4, please check neutron logs for more information. [ 579.374169] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] [ 579.374169] env[62552]: INFO nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Terminating instance [ 579.492977] env[62552]: DEBUG oslo_vmware.api [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Task: {'id': task-1239193, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152939} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.493293] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 579.493481] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 579.493653] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.493837] env[62552]: INFO nova.compute.manager [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Took 1.10 seconds to destroy the instance on the hypervisor. [ 579.494090] env[62552]: DEBUG oslo.service.loopingcall [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.494279] env[62552]: DEBUG nova.compute.manager [-] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 579.494376] env[62552]: DEBUG nova.network.neutron [-] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.534760] env[62552]: DEBUG oslo_vmware.rw_handles [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Completed reading data from the image iterator. {{(pid=62552) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 579.534982] env[62552]: DEBUG oslo_vmware.rw_handles [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 579.554230] env[62552]: DEBUG nova.network.neutron [-] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.585664] env[62552]: DEBUG nova.scheduler.client.report [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 579.594113] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Downloaded image file data 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee to vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk on the data store datastore1 {{(pid=62552) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 579.596159] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Caching image {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 579.596451] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Copying Virtual Disk [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk to [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 579.597420] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04d5ba07-024d-4ea7-92f9-f3067f5fd229 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.605521] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 579.605521] env[62552]: value = "task-1239194" [ 579.605521] env[62552]: _type = "Task" [ 579.605521] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.617761] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239194, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.639969] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Releasing lock "refresh_cache-b3dcf7cf-5807-4483-8ca0-a245931c7dc3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.640259] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 579.640431] env[62552]: DEBUG nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 579.640600] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.693172] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.857748] env[62552]: INFO nova.scheduler.client.report [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Deleted allocations for instance 05687d2d-0760-4c5d-a231-4ef7c2d440e6 [ 579.876400] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Acquiring lock "refresh_cache-fab97d93-2322-469d-84fa-812e1d491030" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.876886] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Acquired lock "refresh_cache-fab97d93-2322-469d-84fa-812e1d491030" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.877371] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.987189] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.987444] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.057573] env[62552]: DEBUG nova.network.neutron [-] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.092785] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.976s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.092785] env[62552]: ERROR nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dea477e3-cfd3-4ba6-8a50-0199829c3f07, please check neutron logs for more information. [ 580.092785] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] Traceback (most recent call last): [ 580.092785] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 580.092785] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self.driver.spawn(context, instance, image_meta, [ 580.092785] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 580.092785] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.092785] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.092785] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] vm_ref = self.build_virtual_machine(instance, [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] for vif in network_info: [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] return self._sync_wrapper(fn, *args, **kwargs) [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self.wait() [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self[:] = self._gt.wait() [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] return self._exit_event.wait() [ 580.093158] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] result = hub.switch() [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] return self.greenlet.switch() [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] result = function(*args, **kwargs) [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] return func(*args, **kwargs) [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] raise e [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] nwinfo = self.network_api.allocate_for_instance( [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.093502] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] created_port_ids = self._update_ports_for_instance( [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] with excutils.save_and_reraise_exception(): [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] self.force_reraise() [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] raise self.value [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] updated_port = self._update_port( [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] _ensure_no_port_binding_failure(port) [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.093879] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] raise exception.PortBindingFailed(port_id=port['id']) [ 580.094291] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] nova.exception.PortBindingFailed: Binding failed for port dea477e3-cfd3-4ba6-8a50-0199829c3f07, please check neutron logs for more information. [ 580.094291] env[62552]: ERROR nova.compute.manager [instance: 694aaace-af82-4eec-b959-2998d608aac1] [ 580.094291] env[62552]: DEBUG nova.compute.utils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Binding failed for port dea477e3-cfd3-4ba6-8a50-0199829c3f07, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 580.100319] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.525s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.100319] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Build of instance 694aaace-af82-4eec-b959-2998d608aac1 was re-scheduled: Binding failed for port dea477e3-cfd3-4ba6-8a50-0199829c3f07, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 580.103978] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 580.103978] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Acquiring lock "refresh_cache-694aaace-af82-4eec-b959-2998d608aac1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.103978] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Acquired lock "refresh_cache-694aaace-af82-4eec-b959-2998d608aac1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.103978] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 580.115539] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239194, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.195743] env[62552]: DEBUG nova.network.neutron [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.367257] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a91a7082-6f75-4865-a8dd-e43a3366e250 tempest-ImagesOneServerTestJSON-1467523071 tempest-ImagesOneServerTestJSON-1467523071-project-member] Lock "05687d2d-0760-4c5d-a231-4ef7c2d440e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.702s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.406588] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.532456] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.563076] env[62552]: INFO nova.compute.manager [-] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Took 1.07 seconds to deallocate network for instance. [ 580.630957] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239194, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.694701} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.633403] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Copied Virtual Disk [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk to [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 580.633403] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Deleting the datastore file [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/tmp-sparse.vmdk {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 580.633403] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17ddf456-6756-4b60-9ba6-023a26d0a1c9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.636793] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.641091] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 580.641091] env[62552]: value = "task-1239195" [ 580.641091] env[62552]: _type = "Task" [ 580.641091] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.653327] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239195, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.703266] env[62552]: INFO nova.compute.manager [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] [instance: b3dcf7cf-5807-4483-8ca0-a245931c7dc3] Took 1.06 seconds to deallocate network for instance. [ 580.711150] env[62552]: DEBUG nova.compute.manager [req-dcb7e70b-0483-40e7-b9db-00fc92b63866 req-9ef6f581-5888-452b-9d1f-f47d521a1be6 service nova] [instance: fab97d93-2322-469d-84fa-812e1d491030] Received event network-changed-6e71a8c2-6b86-47c7-85cf-a31825c39bc4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 580.711514] env[62552]: DEBUG nova.compute.manager [req-dcb7e70b-0483-40e7-b9db-00fc92b63866 req-9ef6f581-5888-452b-9d1f-f47d521a1be6 service nova] [instance: fab97d93-2322-469d-84fa-812e1d491030] Refreshing instance network info cache due to event network-changed-6e71a8c2-6b86-47c7-85cf-a31825c39bc4. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 580.711584] env[62552]: DEBUG oslo_concurrency.lockutils [req-dcb7e70b-0483-40e7-b9db-00fc92b63866 req-9ef6f581-5888-452b-9d1f-f47d521a1be6 service nova] Acquiring lock "refresh_cache-fab97d93-2322-469d-84fa-812e1d491030" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.809481] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.870022] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 580.965086] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681c692f-aa77-44e0-a3e2-3865ee55383f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.974069] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafe4969-ce26-43ff-b7a8-33dd8f823628 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.012465] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12302db-1541-4a8f-9aef-e2b8b25026ad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.020335] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a70909-c156-4260-9b2b-528e889601f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.033685] env[62552]: DEBUG nova.compute.provider_tree [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.036439] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Releasing lock "refresh_cache-fab97d93-2322-469d-84fa-812e1d491030" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.036589] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 581.036776] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 581.037292] env[62552]: DEBUG oslo_concurrency.lockutils [req-dcb7e70b-0483-40e7-b9db-00fc92b63866 req-9ef6f581-5888-452b-9d1f-f47d521a1be6 service nova] Acquired lock "refresh_cache-fab97d93-2322-469d-84fa-812e1d491030" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.037469] env[62552]: DEBUG nova.network.neutron [req-dcb7e70b-0483-40e7-b9db-00fc92b63866 req-9ef6f581-5888-452b-9d1f-f47d521a1be6 service nova] [instance: fab97d93-2322-469d-84fa-812e1d491030] Refreshing network info cache for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 581.038616] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05f08208-9b28-4612-b3cf-a76fb91b25a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.046971] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c11edc9-d667-43bc-8e0e-cb7de398126c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.072625] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fab97d93-2322-469d-84fa-812e1d491030 could not be found. [ 581.072625] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 581.072625] env[62552]: INFO nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Took 0.04 seconds to destroy the instance on the hypervisor. [ 581.072625] env[62552]: DEBUG oslo.service.loopingcall [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.073518] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.074069] env[62552]: DEBUG nova.compute.manager [-] [instance: fab97d93-2322-469d-84fa-812e1d491030] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 581.074069] env[62552]: DEBUG nova.network.neutron [-] [instance: fab97d93-2322-469d-84fa-812e1d491030] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 581.098701] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.099285] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.116944] env[62552]: DEBUG nova.network.neutron [-] [instance: fab97d93-2322-469d-84fa-812e1d491030] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.154429] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239195, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024162} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.154537] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 581.154691] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Moving file from [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee to [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee. {{(pid=62552) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 581.155213] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-8c68949b-d6c6-45e8-9d8d-ea5b7bd47db2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.161815] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 581.161815] env[62552]: value = "task-1239196" [ 581.161815] env[62552]: _type = "Task" [ 581.161815] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.171655] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239196, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.315942] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Releasing lock "refresh_cache-694aaace-af82-4eec-b959-2998d608aac1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.316135] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 581.316340] env[62552]: DEBUG nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 581.316514] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 581.348566] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.405559] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.541021] env[62552]: DEBUG nova.scheduler.client.report [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 581.569909] env[62552]: DEBUG nova.network.neutron [req-dcb7e70b-0483-40e7-b9db-00fc92b63866 req-9ef6f581-5888-452b-9d1f-f47d521a1be6 service nova] [instance: fab97d93-2322-469d-84fa-812e1d491030] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.620908] env[62552]: DEBUG nova.network.neutron [-] [instance: fab97d93-2322-469d-84fa-812e1d491030] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.672825] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239196, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024195} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.673286] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] File moved {{(pid=62552) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 581.673360] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Cleaning up location [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 581.673453] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Deleting the datastore file [datastore1] vmware_temp/72ef307b-df3c-4c76-a8e1-7af4e0ed7bea {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 581.673695] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61ed6ece-3ca2-47cf-bb28-48e0b22e0896 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.681662] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 581.681662] env[62552]: value = "task-1239197" [ 581.681662] env[62552]: _type = "Task" [ 581.681662] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.690100] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.711814] env[62552]: DEBUG nova.network.neutron [req-dcb7e70b-0483-40e7-b9db-00fc92b63866 req-9ef6f581-5888-452b-9d1f-f47d521a1be6 service nova] [instance: fab97d93-2322-469d-84fa-812e1d491030] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.736764] env[62552]: INFO nova.scheduler.client.report [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Deleted allocations for instance b3dcf7cf-5807-4483-8ca0-a245931c7dc3 [ 581.854740] env[62552]: DEBUG nova.network.neutron [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.045482] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.046182] env[62552]: ERROR nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b4978a45-38c6-4be0-97b4-b5d4937c68fe, please check neutron logs for more information. [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Traceback (most recent call last): [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self.driver.spawn(context, instance, image_meta, [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] vm_ref = self.build_virtual_machine(instance, [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.046182] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] for vif in network_info: [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] return self._sync_wrapper(fn, *args, **kwargs) [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self.wait() [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self[:] = self._gt.wait() [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] return self._exit_event.wait() [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] result = hub.switch() [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.046556] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] return self.greenlet.switch() [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] result = function(*args, **kwargs) [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] return func(*args, **kwargs) [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] raise e [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] nwinfo = self.network_api.allocate_for_instance( [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] created_port_ids = self._update_ports_for_instance( [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] with excutils.save_and_reraise_exception(): [ 582.046953] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] self.force_reraise() [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] raise self.value [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] updated_port = self._update_port( [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] _ensure_no_port_binding_failure(port) [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] raise exception.PortBindingFailed(port_id=port['id']) [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] nova.exception.PortBindingFailed: Binding failed for port b4978a45-38c6-4be0-97b4-b5d4937c68fe, please check neutron logs for more information. [ 582.047409] env[62552]: ERROR nova.compute.manager [instance: 1178592c-329f-4af6-94ba-07db03c07f60] [ 582.048316] env[62552]: DEBUG nova.compute.utils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Binding failed for port b4978a45-38c6-4be0-97b4-b5d4937c68fe, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.049502] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Build of instance 1178592c-329f-4af6-94ba-07db03c07f60 was re-scheduled: Binding failed for port b4978a45-38c6-4be0-97b4-b5d4937c68fe, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 582.050098] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 582.051505] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-1178592c-329f-4af6-94ba-07db03c07f60" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.051710] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-1178592c-329f-4af6-94ba-07db03c07f60" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.051910] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 582.054265] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.169s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.123624] env[62552]: INFO nova.compute.manager [-] [instance: fab97d93-2322-469d-84fa-812e1d491030] Took 1.05 seconds to deallocate network for instance. [ 582.126597] env[62552]: DEBUG nova.compute.claims [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 582.126811] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.193243] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239197, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026064} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.193243] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 582.193243] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c15bad6c-6304-477f-8417-443f141e70d4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.198108] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 582.198108] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524ee851-3c96-d396-43f5-5be3cc00d206" [ 582.198108] env[62552]: _type = "Task" [ 582.198108] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.211015] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524ee851-3c96-d396-43f5-5be3cc00d206, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.217275] env[62552]: DEBUG oslo_concurrency.lockutils [req-dcb7e70b-0483-40e7-b9db-00fc92b63866 req-9ef6f581-5888-452b-9d1f-f47d521a1be6 service nova] Releasing lock "refresh_cache-fab97d93-2322-469d-84fa-812e1d491030" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.245075] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1688e000-50c7-491e-8d8b-2f9ee925dccd tempest-ServerExternalEventsTest-1089316746 tempest-ServerExternalEventsTest-1089316746-project-member] Lock "b3dcf7cf-5807-4483-8ca0-a245931c7dc3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.931s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.356226] env[62552]: INFO nova.compute.manager [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] [instance: 694aaace-af82-4eec-b959-2998d608aac1] Took 1.04 seconds to deallocate network for instance. [ 582.605224] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.659710] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "74b58f3f-a575-4790-8a90-cad46b1499e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.660305] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "74b58f3f-a575-4790-8a90-cad46b1499e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.712058] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524ee851-3c96-d396-43f5-5be3cc00d206, 'name': SearchDatastore_Task, 'duration_secs': 0.010122} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.712409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.712691] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 7408a538-6091-4aa2-b2d0-a3d93840b341/7408a538-6091-4aa2-b2d0-a3d93840b341.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 582.713811] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e576273f-ae12-4f6f-a294-d3f7c17ee578 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.721254] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 582.721254] env[62552]: value = "task-1239198" [ 582.721254] env[62552]: _type = "Task" [ 582.721254] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.732932] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239198, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.748985] env[62552]: DEBUG nova.compute.manager [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 582.824676] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.047376] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Acquiring lock "384445d0-93b0-427c-92aa-ef557351a7cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.047376] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Lock "384445d0-93b0-427c-92aa-ef557351a7cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.086694] env[62552]: WARNING nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6240d4e9-79ff-4c84-9d04-49879088bde4 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 583.231524] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239198, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447484} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.231784] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 7408a538-6091-4aa2-b2d0-a3d93840b341/7408a538-6091-4aa2-b2d0-a3d93840b341.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 583.231998] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 583.232252] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8a0b18d-199a-42ee-8ad8-f23db8d7272d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.238224] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 583.238224] env[62552]: value = "task-1239199" [ 583.238224] env[62552]: _type = "Task" [ 583.238224] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.245559] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239199, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.269244] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.328224] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-1178592c-329f-4af6-94ba-07db03c07f60" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.328477] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 583.328659] env[62552]: DEBUG nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 583.328823] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.362029] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.388884] env[62552]: INFO nova.scheduler.client.report [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Deleted allocations for instance 694aaace-af82-4eec-b959-2998d608aac1 [ 583.500813] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "fae607f4-e18a-4cd4-8843-246a3341a548" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.501060] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "fae607f4-e18a-4cd4-8843-246a3341a548" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.591895] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 694aaace-af82-4eec-b959-2998d608aac1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 583.750742] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239199, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064998} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.751075] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 583.752684] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302bd1fa-32c0-4bdd-b8b2-1d65ee876962 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.785882] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] 7408a538-6091-4aa2-b2d0-a3d93840b341/7408a538-6091-4aa2-b2d0-a3d93840b341.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 583.786168] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0333f5c-c945-4a2e-9b4f-4262f07d3812 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.805994] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 583.805994] env[62552]: value = "task-1239200" [ 583.805994] env[62552]: _type = "Task" [ 583.805994] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.814412] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239200, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.866102] env[62552]: DEBUG nova.network.neutron [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.896255] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fd5594b9-2230-4c0a-8975-9d43a22231c8 tempest-ServerMetadataNegativeTestJSON-743061315 tempest-ServerMetadataNegativeTestJSON-743061315-project-member] Lock "694aaace-af82-4eec-b959-2998d608aac1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.921s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.957012] env[62552]: DEBUG nova.compute.manager [req-9718e564-927c-458c-897f-8cafa40999d3 req-7c9ef8b9-8376-45a4-aaa2-26140185c6ef service nova] [instance: fab97d93-2322-469d-84fa-812e1d491030] Received event network-vif-deleted-6e71a8c2-6b86-47c7-85cf-a31825c39bc4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 584.097714] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 1178592c-329f-4af6-94ba-07db03c07f60 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 584.097714] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 7408a538-6091-4aa2-b2d0-a3d93840b341 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 584.097714] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance fab97d93-2322-469d-84fa-812e1d491030 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 584.166179] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "187dd75d-cdb0-4b86-a396-f43778fdcc21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.166433] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "187dd75d-cdb0-4b86-a396-f43778fdcc21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.318392] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239200, 'name': ReconfigVM_Task, 'duration_secs': 0.278711} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.318665] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Reconfigured VM instance instance-0000000a to attach disk [datastore1] 7408a538-6091-4aa2-b2d0-a3d93840b341/7408a538-6091-4aa2-b2d0-a3d93840b341.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 584.319534] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1936710f-8cde-456f-9f26-32e44c7feb0c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.325919] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 584.325919] env[62552]: value = "task-1239201" [ 584.325919] env[62552]: _type = "Task" [ 584.325919] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.334822] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239201, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.368892] env[62552]: INFO nova.compute.manager [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 1178592c-329f-4af6-94ba-07db03c07f60] Took 1.04 seconds to deallocate network for instance. [ 584.399847] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 584.599074] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "1b326e57-365e-49cc-b6f0-eed559ce32f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.599397] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "1b326e57-365e-49cc-b6f0-eed559ce32f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.604766] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance c4a725a7-a14c-44ec-a81a-3e965d7e0817 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 584.840448] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239201, 'name': Rename_Task, 'duration_secs': 0.127921} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.840783] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 584.840965] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-abad14f5-913c-41a0-ad1e-31735916cd3c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.852025] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 584.852025] env[62552]: value = "task-1239202" [ 584.852025] env[62552]: _type = "Task" [ 584.852025] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.861399] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239202, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.938374] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.110309] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance a34ea7fb-bd17-4353-806d-4cd2f5ee169d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 585.363212] env[62552]: DEBUG oslo_vmware.api [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239202, 'name': PowerOnVM_Task, 'duration_secs': 0.44128} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.363212] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 585.363212] env[62552]: DEBUG nova.compute.manager [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 585.363740] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb75f42-2257-4116-8157-2c410fec744c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.422194] env[62552]: INFO nova.scheduler.client.report [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted allocations for instance 1178592c-329f-4af6-94ba-07db03c07f60 [ 585.613186] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance a44de380-895c-4f06-835e-595917a25be1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 585.891193] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.934037] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c4be14b-4422-4c32-9d20-20ace303cae1 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "1178592c-329f-4af6-94ba-07db03c07f60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.857s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.964579] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "90b49142-4fd8-4c7b-98d7-0adaf5038c72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.964579] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "90b49142-4fd8-4c7b-98d7-0adaf5038c72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.120434] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 27e4d3de-3e44-4293-9b04-e511c9f02b42 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 586.438601] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 586.627491] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 92b9d5a7-e5ed-480e-b68e-b589e6d00832 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 586.972582] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.134100] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 58c91f56-a683-4a62-bd45-80c627622621 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 587.444045] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "dd4c1e35-0397-4ee0-932d-2dc7f0870f84" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.444045] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "dd4c1e35-0397-4ee0-932d-2dc7f0870f84" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.494998] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "7408a538-6091-4aa2-b2d0-a3d93840b341" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.495211] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "7408a538-6091-4aa2-b2d0-a3d93840b341" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.496314] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "7408a538-6091-4aa2-b2d0-a3d93840b341-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.496314] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "7408a538-6091-4aa2-b2d0-a3d93840b341-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.496314] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "7408a538-6091-4aa2-b2d0-a3d93840b341-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.504396] env[62552]: INFO nova.compute.manager [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Terminating instance [ 587.638063] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 9952d8f6-ee78-4c2d-b147-8c08c027f440 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.008230] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "refresh_cache-7408a538-6091-4aa2-b2d0-a3d93840b341" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.008230] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquired lock "refresh_cache-7408a538-6091-4aa2-b2d0-a3d93840b341" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.008230] env[62552]: DEBUG nova.network.neutron [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.317336] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 729eba7d-5f14-4506-8fee-24f6eb685b12 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.558146] env[62552]: DEBUG nova.network.neutron [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.572949] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Acquiring lock "4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.572949] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Lock "4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.645443] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance e22c0fef-3d0d-4339-be0b-f129b3628b80 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.719092] env[62552]: DEBUG nova.network.neutron [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.148432] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 589.195455] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.195689] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.212150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "04289f4b-4342-4377-8f9c-3013eb646cc6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.212863] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "04289f4b-4342-4377-8f9c-3013eb646cc6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.224256] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Releasing lock "refresh_cache-7408a538-6091-4aa2-b2d0-a3d93840b341" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.224256] env[62552]: DEBUG nova.compute.manager [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 589.224333] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.226212] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acbf5aa-d2f5-4ebd-b4b8-6a4844c8de03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.236579] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 589.236937] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e33c6b4b-712f-413a-95df-87f274097dc1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.243503] env[62552]: DEBUG oslo_vmware.api [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 589.243503] env[62552]: value = "task-1239203" [ 589.243503] env[62552]: _type = "Task" [ 589.243503] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.261234] env[62552]: DEBUG oslo_vmware.api [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239203, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.264306] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "f8f37b3c-5e74-4459-9177-aacfac707616" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.264630] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "f8f37b3c-5e74-4459-9177-aacfac707616" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.652073] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance b6bff07b-f316-4f26-9623-d50328df1ac0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 589.758862] env[62552]: DEBUG oslo_vmware.api [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239203, 'name': PowerOffVM_Task, 'duration_secs': 0.167117} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.759186] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 589.759421] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 589.759693] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4757215-e90e-4a68-bdf1-6e295570066e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.786365] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 589.786601] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 589.786783] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Deleting the datastore file [datastore1] 7408a538-6091-4aa2-b2d0-a3d93840b341 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 589.787069] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b81a14a-b03f-41dd-afc4-2daba4573500 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.794467] env[62552]: DEBUG oslo_vmware.api [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for the task: (returnval){ [ 589.794467] env[62552]: value = "task-1239205" [ 589.794467] env[62552]: _type = "Task" [ 589.794467] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.805182] env[62552]: DEBUG oslo_vmware.api [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239205, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.159610] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 590.306055] env[62552]: DEBUG oslo_vmware.api [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Task: {'id': task-1239205, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109619} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.306898] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 590.307255] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 590.307944] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.307944] env[62552]: INFO nova.compute.manager [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Took 1.08 seconds to destroy the instance on the hypervisor. [ 590.309893] env[62552]: DEBUG oslo.service.loopingcall [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.309893] env[62552]: DEBUG nova.compute.manager [-] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 590.309893] env[62552]: DEBUG nova.network.neutron [-] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.339207] env[62552]: DEBUG nova.network.neutron [-] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.663519] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 590.844910] env[62552]: DEBUG nova.network.neutron [-] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.167929] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 74b58f3f-a575-4790-8a90-cad46b1499e6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 591.274026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Acquiring lock "2031fc3f-2c79-4f6e-b7eb-29235814141d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.274026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Lock "2031fc3f-2c79-4f6e-b7eb-29235814141d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.350615] env[62552]: INFO nova.compute.manager [-] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Took 1.04 seconds to deallocate network for instance. [ 591.672388] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 384445d0-93b0-427c-92aa-ef557351a7cd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 591.858486] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.158319] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "cf0dcba7-a36e-435c-8c34-d352b49a723b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.158419] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "cf0dcba7-a36e-435c-8c34-d352b49a723b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.174935] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance fae607f4-e18a-4cd4-8843-246a3341a548 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 592.176143] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 592.176143] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 592.682996] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feccfcc9-fa30-4844-bcea-b01360d01921 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.691379] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db006e0-b17f-44c6-94c2-7eb93932d7f0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.726318] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2345d2ae-5242-4adf-a032-c82556316bfe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.734514] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a313031-d3c5-43d2-b21a-12c4e0332aff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.748212] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.253602] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 593.762020] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 593.762020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.707s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.762020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.994s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.763322] env[62552]: INFO nova.compute.claims [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.269513] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fcb80c-1f9b-4dc0-b636-27515e447087 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.280343] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79dfb787-06e7-43d9-8146-63a283e04166 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.315314] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0038cd5f-4f36-4f89-90e8-310c58f84991 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.324791] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89c92c7-9ba7-4ac2-9ee8-ed0c3da43b41 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.340122] env[62552]: DEBUG nova.compute.provider_tree [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.843491] env[62552]: DEBUG nova.scheduler.client.report [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 596.354455] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.592s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.354988] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 596.361169] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.315s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.369049] env[62552]: INFO nova.compute.claims [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.873433] env[62552]: DEBUG nova.compute.utils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.880474] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 596.880474] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 597.206877] env[62552]: DEBUG nova.policy [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee6f13726f4a4cf9a9bf660606de018b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a52552bd7c184d3a97a1f27f658b0ea9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 597.380011] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 597.863204] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5785318b-efb0-40a4-96d9-807bbfe7104b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.878040] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c3468e-13c6-4c0b-aab4-658478c6a823 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.922582] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe88f7f-4089-4762-9f58-d704bd7765d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.927977] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beaafa1e-a00a-4e8a-ad4d-7cabdbebf256 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.945514] env[62552]: DEBUG nova.compute.provider_tree [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.089464] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Acquiring lock "a9a16b57-603a-4d43-b834-e6ea62675f3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.089728] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Lock "a9a16b57-603a-4d43-b834-e6ea62675f3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.307575] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Successfully created port: e5b1d0f8-689f-4a7f-99a1-52883e53bfa5 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.425150] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 598.450997] env[62552]: DEBUG nova.scheduler.client.report [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 598.465974] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.466271] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.466419] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.466645] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.466792] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.466938] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.467163] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.467339] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.467500] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.467676] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.467855] env[62552]: DEBUG nova.virt.hardware [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.468778] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20b0000-0619-4977-97f2-9be4859fdf4c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.476908] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "5ae752f6-9fd4-46fd-a7d2-558af22e2943" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.476908] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "5ae752f6-9fd4-46fd-a7d2-558af22e2943" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.485253] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef47144-a8fd-4267-b75e-dee4d48677d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.967248] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.967248] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 598.970950] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.923s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.970950] env[62552]: INFO nova.compute.claims [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.480353] env[62552]: DEBUG nova.compute.utils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.484614] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 599.484777] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 599.836475] env[62552]: DEBUG nova.policy [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8dfbb8b7e4c044b89ad167f40190e1c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '57f118c6cb224e87bf92a860c386a6af', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 599.993127] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 600.508891] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b67eb70-a13d-48f6-8b3d-64990b6b3f6e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.516889] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ce2390-e05b-46f5-9eb7-fada7055969e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.553897] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c59cc99-cbac-47bd-8ff8-4694a741861a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.562433] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf6d02a-30c9-4575-b497-e0828e6336a9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.577412] env[62552]: DEBUG nova.compute.provider_tree [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.669742] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Successfully created port: 74e5cee9-1240-45ee-84b4-f9afe2710f71 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.002727] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 601.037477] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.037778] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.038114] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.038333] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.038578] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.038764] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.038899] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.039273] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.039490] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.039664] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.039839] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.040721] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d305e8-c2fa-4ec2-bb0f-890eae08803c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.048984] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355baf80-2781-4003-bbe1-a1119f74cf02 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.080758] env[62552]: DEBUG nova.scheduler.client.report [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 601.589291] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.589291] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 601.590533] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.766s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.591954] env[62552]: INFO nova.compute.claims [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 601.930191] env[62552]: ERROR nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5, please check neutron logs for more information. [ 601.930191] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.930191] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.930191] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.930191] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.930191] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.930191] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.930191] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.930191] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.930191] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 601.930191] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.930191] env[62552]: ERROR nova.compute.manager raise self.value [ 601.930191] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.930191] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.930191] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.930191] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.930905] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.930905] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.930905] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5, please check neutron logs for more information. [ 601.930905] env[62552]: ERROR nova.compute.manager [ 601.930905] env[62552]: Traceback (most recent call last): [ 601.930905] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.930905] env[62552]: listener.cb(fileno) [ 601.930905] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.930905] env[62552]: result = function(*args, **kwargs) [ 601.930905] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.930905] env[62552]: return func(*args, **kwargs) [ 601.930905] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.930905] env[62552]: raise e [ 601.930905] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.930905] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 601.930905] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.930905] env[62552]: created_port_ids = self._update_ports_for_instance( [ 601.930905] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.930905] env[62552]: with excutils.save_and_reraise_exception(): [ 601.930905] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.930905] env[62552]: self.force_reraise() [ 601.930905] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.930905] env[62552]: raise self.value [ 601.930905] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.930905] env[62552]: updated_port = self._update_port( [ 601.930905] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.930905] env[62552]: _ensure_no_port_binding_failure(port) [ 601.930905] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.930905] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.932045] env[62552]: nova.exception.PortBindingFailed: Binding failed for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5, please check neutron logs for more information. [ 601.932045] env[62552]: Removing descriptor: 19 [ 601.932045] env[62552]: ERROR nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5, please check neutron logs for more information. [ 601.932045] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Traceback (most recent call last): [ 601.932045] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 601.932045] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] yield resources [ 601.932045] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 601.932045] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self.driver.spawn(context, instance, image_meta, [ 601.932045] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 601.932045] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.932045] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.932045] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] vm_ref = self.build_virtual_machine(instance, [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] for vif in network_info: [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] return self._sync_wrapper(fn, *args, **kwargs) [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self.wait() [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self[:] = self._gt.wait() [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] return self._exit_event.wait() [ 601.934048] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] result = hub.switch() [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] return self.greenlet.switch() [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] result = function(*args, **kwargs) [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] return func(*args, **kwargs) [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] raise e [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] nwinfo = self.network_api.allocate_for_instance( [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.934493] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] created_port_ids = self._update_ports_for_instance( [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] with excutils.save_and_reraise_exception(): [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self.force_reraise() [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] raise self.value [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] updated_port = self._update_port( [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] _ensure_no_port_binding_failure(port) [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.934932] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] raise exception.PortBindingFailed(port_id=port['id']) [ 601.935680] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] nova.exception.PortBindingFailed: Binding failed for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5, please check neutron logs for more information. [ 601.935680] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] [ 601.935680] env[62552]: INFO nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Terminating instance [ 601.955141] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Acquiring lock "cfdee9ce-8c90-40cc-a09e-2386c719c02e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.955795] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Lock "cfdee9ce-8c90-40cc-a09e-2386c719c02e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.096405] env[62552]: DEBUG nova.compute.utils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.101777] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 602.101777] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 602.220584] env[62552]: DEBUG nova.policy [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8dfbb8b7e4c044b89ad167f40190e1c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '57f118c6cb224e87bf92a860c386a6af', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 602.441824] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "refresh_cache-c4a725a7-a14c-44ec-a81a-3e965d7e0817" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.441990] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquired lock "refresh_cache-c4a725a7-a14c-44ec-a81a-3e965d7e0817" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.442180] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.557532] env[62552]: DEBUG nova.compute.manager [req-fe19e433-6632-46bb-8ceb-c796d88b5303 req-f56509a5-20dc-49c1-b953-23d6b49d92aa service nova] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Received event network-changed-e5b1d0f8-689f-4a7f-99a1-52883e53bfa5 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 602.557532] env[62552]: DEBUG nova.compute.manager [req-fe19e433-6632-46bb-8ceb-c796d88b5303 req-f56509a5-20dc-49c1-b953-23d6b49d92aa service nova] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Refreshing instance network info cache due to event network-changed-e5b1d0f8-689f-4a7f-99a1-52883e53bfa5. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 602.557532] env[62552]: DEBUG oslo_concurrency.lockutils [req-fe19e433-6632-46bb-8ceb-c796d88b5303 req-f56509a5-20dc-49c1-b953-23d6b49d92aa service nova] Acquiring lock "refresh_cache-c4a725a7-a14c-44ec-a81a-3e965d7e0817" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.599936] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 602.989133] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.034827] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Successfully created port: c059d502-667b-4152-bc99-bf65545180f7 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.079401] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Acquiring lock "e8ec980c-0186-48cb-84b9-af902f23d2fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.080086] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Lock "e8ec980c-0186-48cb-84b9-af902f23d2fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.178063] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ffd041-1ce4-432f-a34a-082187a08f54 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.186795] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89b8b3c-86a6-4113-a170-dea70bf2e050 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.219133] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550821c6-ac5a-4312-98d3-88d390896f41 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.226753] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b43296-b419-45fb-96bc-cb7475763494 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.240358] env[62552]: DEBUG nova.compute.provider_tree [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.266663] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.613200] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 603.646560] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 603.646560] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 603.648054] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 603.648054] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 603.648054] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 603.648054] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 603.648054] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 603.648310] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 603.648310] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 603.648310] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 603.648438] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 603.649275] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30f261c-8998-4fcb-9383-49924ae25816 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.659340] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fa78d6-d0fb-454a-bf15-a616278d757d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.743376] env[62552]: DEBUG nova.scheduler.client.report [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 603.762410] env[62552]: ERROR nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 74e5cee9-1240-45ee-84b4-f9afe2710f71, please check neutron logs for more information. [ 603.762410] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 603.762410] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 603.762410] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 603.762410] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.762410] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 603.762410] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.762410] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 603.762410] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.762410] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 603.762410] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.762410] env[62552]: ERROR nova.compute.manager raise self.value [ 603.762410] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.762410] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 603.762410] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.762410] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 603.762948] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.762948] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 603.762948] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 74e5cee9-1240-45ee-84b4-f9afe2710f71, please check neutron logs for more information. [ 603.762948] env[62552]: ERROR nova.compute.manager [ 603.762948] env[62552]: Traceback (most recent call last): [ 603.762948] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 603.762948] env[62552]: listener.cb(fileno) [ 603.762948] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.762948] env[62552]: result = function(*args, **kwargs) [ 603.762948] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.762948] env[62552]: return func(*args, **kwargs) [ 603.762948] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 603.762948] env[62552]: raise e [ 603.762948] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 603.762948] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 603.762948] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.762948] env[62552]: created_port_ids = self._update_ports_for_instance( [ 603.762948] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.762948] env[62552]: with excutils.save_and_reraise_exception(): [ 603.762948] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.762948] env[62552]: self.force_reraise() [ 603.762948] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.762948] env[62552]: raise self.value [ 603.762948] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.762948] env[62552]: updated_port = self._update_port( [ 603.762948] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.762948] env[62552]: _ensure_no_port_binding_failure(port) [ 603.762948] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.762948] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 603.763839] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 74e5cee9-1240-45ee-84b4-f9afe2710f71, please check neutron logs for more information. [ 603.763839] env[62552]: Removing descriptor: 16 [ 603.764241] env[62552]: ERROR nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 74e5cee9-1240-45ee-84b4-f9afe2710f71, please check neutron logs for more information. [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Traceback (most recent call last): [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] yield resources [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self.driver.spawn(context, instance, image_meta, [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] vm_ref = self.build_virtual_machine(instance, [ 603.764241] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] for vif in network_info: [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] return self._sync_wrapper(fn, *args, **kwargs) [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self.wait() [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self[:] = self._gt.wait() [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] return self._exit_event.wait() [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.764618] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] result = hub.switch() [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] return self.greenlet.switch() [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] result = function(*args, **kwargs) [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] return func(*args, **kwargs) [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] raise e [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] nwinfo = self.network_api.allocate_for_instance( [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] created_port_ids = self._update_ports_for_instance( [ 603.765099] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] with excutils.save_and_reraise_exception(): [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self.force_reraise() [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] raise self.value [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] updated_port = self._update_port( [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] _ensure_no_port_binding_failure(port) [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] raise exception.PortBindingFailed(port_id=port['id']) [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] nova.exception.PortBindingFailed: Binding failed for port 74e5cee9-1240-45ee-84b4-f9afe2710f71, please check neutron logs for more information. [ 603.767042] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] [ 603.767470] env[62552]: INFO nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Terminating instance [ 603.771060] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Releasing lock "refresh_cache-c4a725a7-a14c-44ec-a81a-3e965d7e0817" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.771060] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 603.771060] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.771223] env[62552]: DEBUG oslo_concurrency.lockutils [req-fe19e433-6632-46bb-8ceb-c796d88b5303 req-f56509a5-20dc-49c1-b953-23d6b49d92aa service nova] Acquired lock "refresh_cache-c4a725a7-a14c-44ec-a81a-3e965d7e0817" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.771411] env[62552]: DEBUG nova.network.neutron [req-fe19e433-6632-46bb-8ceb-c796d88b5303 req-f56509a5-20dc-49c1-b953-23d6b49d92aa service nova] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Refreshing network info cache for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.772209] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ae0cb10-e0b6-4f61-b8c9-8a73ff3e6238 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.788543] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924fd307-5080-46a5-b013-5edec3eaf51e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.811739] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c4a725a7-a14c-44ec-a81a-3e965d7e0817 could not be found. [ 603.811953] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.812553] env[62552]: INFO nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.812832] env[62552]: DEBUG oslo.service.loopingcall [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.813055] env[62552]: DEBUG nova.compute.manager [-] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 603.813150] env[62552]: DEBUG nova.network.neutron [-] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.834782] env[62552]: DEBUG nova.network.neutron [-] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.252293] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.661s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.252293] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 604.255527] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.910s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.257154] env[62552]: INFO nova.compute.claims [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.279400] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "refresh_cache-27e4d3de-3e44-4293-9b04-e511c9f02b42" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.279606] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquired lock "refresh_cache-27e4d3de-3e44-4293-9b04-e511c9f02b42" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.279782] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.337357] env[62552]: DEBUG nova.network.neutron [-] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.340880] env[62552]: DEBUG nova.network.neutron [req-fe19e433-6632-46bb-8ceb-c796d88b5303 req-f56509a5-20dc-49c1-b953-23d6b49d92aa service nova] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.467822] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Acquiring lock "ab363b52-ade2-45c4-ba3a-ae46ffaf831e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.468076] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Lock "ab363b52-ade2-45c4-ba3a-ae46ffaf831e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.570989] env[62552]: DEBUG nova.network.neutron [req-fe19e433-6632-46bb-8ceb-c796d88b5303 req-f56509a5-20dc-49c1-b953-23d6b49d92aa service nova] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.761903] env[62552]: DEBUG nova.compute.utils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 604.767464] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 604.767707] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 604.803393] env[62552]: DEBUG nova.compute.manager [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Received event network-vif-deleted-e5b1d0f8-689f-4a7f-99a1-52883e53bfa5 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 604.805395] env[62552]: DEBUG nova.compute.manager [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Received event network-changed-74e5cee9-1240-45ee-84b4-f9afe2710f71 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 604.805395] env[62552]: DEBUG nova.compute.manager [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Refreshing instance network info cache due to event network-changed-74e5cee9-1240-45ee-84b4-f9afe2710f71. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 604.805395] env[62552]: DEBUG oslo_concurrency.lockutils [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] Acquiring lock "refresh_cache-27e4d3de-3e44-4293-9b04-e511c9f02b42" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.845229] env[62552]: INFO nova.compute.manager [-] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Took 1.03 seconds to deallocate network for instance. [ 604.848685] env[62552]: DEBUG nova.compute.claims [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.849711] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.856715] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.951691] env[62552]: DEBUG nova.policy [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8dfbb8b7e4c044b89ad167f40190e1c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '57f118c6cb224e87bf92a860c386a6af', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.075597] env[62552]: DEBUG oslo_concurrency.lockutils [req-fe19e433-6632-46bb-8ceb-c796d88b5303 req-f56509a5-20dc-49c1-b953-23d6b49d92aa service nova] Releasing lock "refresh_cache-c4a725a7-a14c-44ec-a81a-3e965d7e0817" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.128242] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.268527] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 605.335343] env[62552]: ERROR nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. [ 605.335343] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.335343] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.335343] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.335343] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.335343] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.335343] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.335343] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.335343] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.335343] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 605.335343] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.335343] env[62552]: ERROR nova.compute.manager raise self.value [ 605.335343] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.335343] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.335343] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.335343] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.335886] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.335886] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.335886] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. [ 605.335886] env[62552]: ERROR nova.compute.manager [ 605.335886] env[62552]: Traceback (most recent call last): [ 605.335886] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.335886] env[62552]: listener.cb(fileno) [ 605.335886] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.335886] env[62552]: result = function(*args, **kwargs) [ 605.335886] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.335886] env[62552]: return func(*args, **kwargs) [ 605.335886] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.335886] env[62552]: raise e [ 605.335886] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.335886] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 605.335886] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.335886] env[62552]: created_port_ids = self._update_ports_for_instance( [ 605.335886] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.335886] env[62552]: with excutils.save_and_reraise_exception(): [ 605.335886] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.335886] env[62552]: self.force_reraise() [ 605.335886] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.335886] env[62552]: raise self.value [ 605.335886] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.335886] env[62552]: updated_port = self._update_port( [ 605.335886] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.335886] env[62552]: _ensure_no_port_binding_failure(port) [ 605.335886] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.335886] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.336704] env[62552]: nova.exception.PortBindingFailed: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. [ 605.336704] env[62552]: Removing descriptor: 19 [ 605.338174] env[62552]: ERROR nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] Traceback (most recent call last): [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] yield resources [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self.driver.spawn(context, instance, image_meta, [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] vm_ref = self.build_virtual_machine(instance, [ 605.338174] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] for vif in network_info: [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] return self._sync_wrapper(fn, *args, **kwargs) [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self.wait() [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self[:] = self._gt.wait() [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] return self._exit_event.wait() [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.338882] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] result = hub.switch() [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] return self.greenlet.switch() [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] result = function(*args, **kwargs) [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] return func(*args, **kwargs) [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] raise e [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] nwinfo = self.network_api.allocate_for_instance( [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] created_port_ids = self._update_ports_for_instance( [ 605.340254] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] with excutils.save_and_reraise_exception(): [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self.force_reraise() [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] raise self.value [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] updated_port = self._update_port( [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] _ensure_no_port_binding_failure(port) [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] raise exception.PortBindingFailed(port_id=port['id']) [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] nova.exception.PortBindingFailed: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. [ 605.341257] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] [ 605.341941] env[62552]: INFO nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Terminating instance [ 605.633040] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Releasing lock "refresh_cache-27e4d3de-3e44-4293-9b04-e511c9f02b42" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.633040] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 605.633040] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 605.633361] env[62552]: DEBUG oslo_concurrency.lockutils [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] Acquired lock "refresh_cache-27e4d3de-3e44-4293-9b04-e511c9f02b42" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.633555] env[62552]: DEBUG nova.network.neutron [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Refreshing network info cache for port 74e5cee9-1240-45ee-84b4-f9afe2710f71 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 605.636962] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4475584a-d119-4786-ad5e-c3ccb5aef5bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.647382] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8ec7ed-379d-4de7-975e-d5298d4fc490 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.677522] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 27e4d3de-3e44-4293-9b04-e511c9f02b42 could not be found. [ 605.677877] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 605.678152] env[62552]: INFO nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Took 0.05 seconds to destroy the instance on the hypervisor. [ 605.680238] env[62552]: DEBUG oslo.service.loopingcall [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.680238] env[62552]: DEBUG nova.compute.manager [-] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 605.680238] env[62552]: DEBUG nova.network.neutron [-] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.692595] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Successfully created port: 673852d0-7384-41cc-912f-b25d6edbb3d1 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 605.731478] env[62552]: DEBUG nova.network.neutron [-] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.841973] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "refresh_cache-a44de380-895c-4f06-835e-595917a25be1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.846600] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquired lock "refresh_cache-a44de380-895c-4f06-835e-595917a25be1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.846600] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.883629] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03716d11-ea0f-45bd-b7c1-4b0a77c69645 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.894056] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f166357d-2e5a-4f13-9872-ec621e8105ed {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.929112] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089f1f7a-b4fc-46b2-9d70-7d83e9264f19 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.940782] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf4cfa8-4656-493b-9791-344b08fe2946 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.958891] env[62552]: DEBUG nova.compute.provider_tree [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.238838] env[62552]: DEBUG nova.network.neutron [-] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.286330] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 606.315500] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.315500] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.315500] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.315730] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.316092] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.321017] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.321017] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.321017] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.321017] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.321017] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.321278] env[62552]: DEBUG nova.virt.hardware [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.321278] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3b521e-da0b-491e-a80c-fe63e9edb912 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.330880] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0276d131-760d-4908-aada-5bd5d34dc8c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.336905] env[62552]: DEBUG nova.network.neutron [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.390454] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.463384] env[62552]: DEBUG nova.scheduler.client.report [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 606.568629] env[62552]: DEBUG nova.network.neutron [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.612832] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.742353] env[62552]: INFO nova.compute.manager [-] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Took 1.06 seconds to deallocate network for instance. [ 606.745131] env[62552]: DEBUG nova.compute.claims [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 606.746672] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.967492] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.712s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.968025] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 606.971482] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.898s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.971720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.973868] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.568s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.975223] env[62552]: INFO nova.compute.claims [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.014244] env[62552]: INFO nova.scheduler.client.report [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Deleted allocations for instance 6240d4e9-79ff-4c84-9d04-49879088bde4 [ 607.072736] env[62552]: DEBUG oslo_concurrency.lockutils [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] Releasing lock "refresh_cache-27e4d3de-3e44-4293-9b04-e511c9f02b42" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.074938] env[62552]: DEBUG nova.compute.manager [req-89572d19-30e8-4320-a971-b23d61571d0e req-8fcad0b4-edc1-437f-b113-a8c8b3140951 service nova] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Received event network-vif-deleted-74e5cee9-1240-45ee-84b4-f9afe2710f71 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 607.085928] env[62552]: DEBUG nova.compute.manager [req-6e9d2128-3346-4a11-aedd-20a63f389c37 req-765c1f94-f530-4c40-8a17-388c7302dffb service nova] [instance: a44de380-895c-4f06-835e-595917a25be1] Received event network-changed-c059d502-667b-4152-bc99-bf65545180f7 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 607.086106] env[62552]: DEBUG nova.compute.manager [req-6e9d2128-3346-4a11-aedd-20a63f389c37 req-765c1f94-f530-4c40-8a17-388c7302dffb service nova] [instance: a44de380-895c-4f06-835e-595917a25be1] Refreshing instance network info cache due to event network-changed-c059d502-667b-4152-bc99-bf65545180f7. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 607.087949] env[62552]: DEBUG oslo_concurrency.lockutils [req-6e9d2128-3346-4a11-aedd-20a63f389c37 req-765c1f94-f530-4c40-8a17-388c7302dffb service nova] Acquiring lock "refresh_cache-a44de380-895c-4f06-835e-595917a25be1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.116979] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Releasing lock "refresh_cache-a44de380-895c-4f06-835e-595917a25be1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.116979] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 607.116979] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.116979] env[62552]: DEBUG oslo_concurrency.lockutils [req-6e9d2128-3346-4a11-aedd-20a63f389c37 req-765c1f94-f530-4c40-8a17-388c7302dffb service nova] Acquired lock "refresh_cache-a44de380-895c-4f06-835e-595917a25be1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.116979] env[62552]: DEBUG nova.network.neutron [req-6e9d2128-3346-4a11-aedd-20a63f389c37 req-765c1f94-f530-4c40-8a17-388c7302dffb service nova] [instance: a44de380-895c-4f06-835e-595917a25be1] Refreshing network info cache for port c059d502-667b-4152-bc99-bf65545180f7 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 607.117927] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51d43afa-5857-4b63-b483-7cc7667fae7d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.128774] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133cf56f-ce2c-4d4c-90fb-e3355ddbdcf9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.153344] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a44de380-895c-4f06-835e-595917a25be1 could not be found. [ 607.156025] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.156025] env[62552]: INFO nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 607.156025] env[62552]: DEBUG oslo.service.loopingcall [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.156025] env[62552]: DEBUG nova.compute.manager [-] [instance: a44de380-895c-4f06-835e-595917a25be1] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 607.156025] env[62552]: DEBUG nova.network.neutron [-] [instance: a44de380-895c-4f06-835e-595917a25be1] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.174814] env[62552]: DEBUG nova.network.neutron [-] [instance: a44de380-895c-4f06-835e-595917a25be1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.480797] env[62552]: DEBUG nova.compute.utils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.486636] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 607.487850] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 607.529021] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0adf5c01-712e-4829-89b0-c50e272d55ed tempest-ServerDiagnosticsV248Test-1882194153 tempest-ServerDiagnosticsV248Test-1882194153-project-member] Lock "6240d4e9-79ff-4c84-9d04-49879088bde4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.959s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.600142] env[62552]: DEBUG nova.policy [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4aa9d155dc624ad69f0acc36dbf04f6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f000fc33f694103ab3cb7d5bb09cc3f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.670460] env[62552]: DEBUG nova.network.neutron [req-6e9d2128-3346-4a11-aedd-20a63f389c37 req-765c1f94-f530-4c40-8a17-388c7302dffb service nova] [instance: a44de380-895c-4f06-835e-595917a25be1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.677437] env[62552]: DEBUG nova.network.neutron [-] [instance: a44de380-895c-4f06-835e-595917a25be1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.889961] env[62552]: ERROR nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. [ 607.889961] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.889961] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 607.889961] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.889961] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.889961] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.889961] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.889961] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.889961] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.889961] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 607.889961] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.889961] env[62552]: ERROR nova.compute.manager raise self.value [ 607.889961] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.889961] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.889961] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.889961] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.890492] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.890492] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.890492] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. [ 607.890492] env[62552]: ERROR nova.compute.manager [ 607.890492] env[62552]: Traceback (most recent call last): [ 607.890492] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.890492] env[62552]: listener.cb(fileno) [ 607.890492] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.890492] env[62552]: result = function(*args, **kwargs) [ 607.890492] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.890492] env[62552]: return func(*args, **kwargs) [ 607.890492] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 607.890492] env[62552]: raise e [ 607.890492] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 607.890492] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 607.890492] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.890492] env[62552]: created_port_ids = self._update_ports_for_instance( [ 607.890492] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.890492] env[62552]: with excutils.save_and_reraise_exception(): [ 607.890492] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.890492] env[62552]: self.force_reraise() [ 607.890492] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.890492] env[62552]: raise self.value [ 607.890492] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.890492] env[62552]: updated_port = self._update_port( [ 607.890492] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.890492] env[62552]: _ensure_no_port_binding_failure(port) [ 607.890492] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.890492] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.891302] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. [ 607.891302] env[62552]: Removing descriptor: 16 [ 607.891302] env[62552]: ERROR nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. [ 607.891302] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Traceback (most recent call last): [ 607.891302] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 607.891302] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] yield resources [ 607.891302] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 607.891302] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self.driver.spawn(context, instance, image_meta, [ 607.891302] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 607.891302] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.891302] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.891302] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] vm_ref = self.build_virtual_machine(instance, [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] for vif in network_info: [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] return self._sync_wrapper(fn, *args, **kwargs) [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self.wait() [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self[:] = self._gt.wait() [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] return self._exit_event.wait() [ 607.891653] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] result = hub.switch() [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] return self.greenlet.switch() [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] result = function(*args, **kwargs) [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] return func(*args, **kwargs) [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] raise e [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] nwinfo = self.network_api.allocate_for_instance( [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.892141] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] created_port_ids = self._update_ports_for_instance( [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] with excutils.save_and_reraise_exception(): [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self.force_reraise() [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] raise self.value [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] updated_port = self._update_port( [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] _ensure_no_port_binding_failure(port) [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.892558] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] raise exception.PortBindingFailed(port_id=port['id']) [ 607.892968] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] nova.exception.PortBindingFailed: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. [ 607.892968] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] [ 607.892968] env[62552]: INFO nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Terminating instance [ 607.894669] env[62552]: DEBUG nova.network.neutron [req-6e9d2128-3346-4a11-aedd-20a63f389c37 req-765c1f94-f530-4c40-8a17-388c7302dffb service nova] [instance: a44de380-895c-4f06-835e-595917a25be1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.990019] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 608.182131] env[62552]: INFO nova.compute.manager [-] [instance: a44de380-895c-4f06-835e-595917a25be1] Took 1.03 seconds to deallocate network for instance. [ 608.188258] env[62552]: DEBUG nova.compute.claims [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.188502] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.301156] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Successfully created port: 37c536bd-77ca-4c85-b9cb-65a824f863fe {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.407729] env[62552]: DEBUG oslo_concurrency.lockutils [req-6e9d2128-3346-4a11-aedd-20a63f389c37 req-765c1f94-f530-4c40-8a17-388c7302dffb service nova] Releasing lock "refresh_cache-a44de380-895c-4f06-835e-595917a25be1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.407729] env[62552]: DEBUG nova.compute.manager [req-6e9d2128-3346-4a11-aedd-20a63f389c37 req-765c1f94-f530-4c40-8a17-388c7302dffb service nova] [instance: a44de380-895c-4f06-835e-595917a25be1] Received event network-vif-deleted-c059d502-667b-4152-bc99-bf65545180f7 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 608.407729] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "refresh_cache-92b9d5a7-e5ed-480e-b68e-b589e6d00832" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.407729] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquired lock "refresh_cache-92b9d5a7-e5ed-480e-b68e-b589e6d00832" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.407729] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.526023] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a8b710-9c0d-44b0-bdb4-ac7181525972 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.535194] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47a9c8b-8e8f-4431-837d-531f64432b42 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.570577] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999bfbdf-e9f0-45e0-a625-a08c73da8e9d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.578823] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8ec036-a70d-491e-b92f-fc36ba1d8d8b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.596762] env[62552]: DEBUG nova.compute.provider_tree [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.950989] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.007018] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 609.044030] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.044030] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.044030] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.044299] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.044299] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.044299] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.044917] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.045230] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.045672] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.046034] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.046361] env[62552]: DEBUG nova.virt.hardware [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.047421] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53bf483b-5616-45c2-8153-554b4b6a1d5a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.057018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051a70ff-733c-42b8-8f1e-085d1d99e420 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.100181] env[62552]: DEBUG nova.scheduler.client.report [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 609.217438] env[62552]: DEBUG nova.compute.manager [req-cd990b5a-12b5-48bd-b6fb-9e255d971fb2 req-09515146-2bb3-4b8a-abf5-d37152fe13dc service nova] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Received event network-changed-673852d0-7384-41cc-912f-b25d6edbb3d1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 609.217438] env[62552]: DEBUG nova.compute.manager [req-cd990b5a-12b5-48bd-b6fb-9e255d971fb2 req-09515146-2bb3-4b8a-abf5-d37152fe13dc service nova] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Refreshing instance network info cache due to event network-changed-673852d0-7384-41cc-912f-b25d6edbb3d1. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 609.217856] env[62552]: DEBUG oslo_concurrency.lockutils [req-cd990b5a-12b5-48bd-b6fb-9e255d971fb2 req-09515146-2bb3-4b8a-abf5-d37152fe13dc service nova] Acquiring lock "refresh_cache-92b9d5a7-e5ed-480e-b68e-b589e6d00832" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.252854] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.610237] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.636s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.610810] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 609.613980] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.487s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.755850] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Releasing lock "refresh_cache-92b9d5a7-e5ed-480e-b68e-b589e6d00832" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.756321] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 609.756545] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 609.757118] env[62552]: DEBUG oslo_concurrency.lockutils [req-cd990b5a-12b5-48bd-b6fb-9e255d971fb2 req-09515146-2bb3-4b8a-abf5-d37152fe13dc service nova] Acquired lock "refresh_cache-92b9d5a7-e5ed-480e-b68e-b589e6d00832" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.757118] env[62552]: DEBUG nova.network.neutron [req-cd990b5a-12b5-48bd-b6fb-9e255d971fb2 req-09515146-2bb3-4b8a-abf5-d37152fe13dc service nova] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Refreshing network info cache for port 673852d0-7384-41cc-912f-b25d6edbb3d1 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 609.761258] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca2ef1fd-b626-4b0e-9191-f794a20312bf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.774039] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82be16f3-b328-4e70-809d-df2a1836c404 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.802971] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92b9d5a7-e5ed-480e-b68e-b589e6d00832 could not be found. [ 609.803471] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 609.803831] env[62552]: INFO nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Took 0.05 seconds to destroy the instance on the hypervisor. [ 609.804272] env[62552]: DEBUG oslo.service.loopingcall [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.804733] env[62552]: DEBUG nova.compute.manager [-] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 609.805000] env[62552]: DEBUG nova.network.neutron [-] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 609.852902] env[62552]: DEBUG nova.network.neutron [-] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.120948] env[62552]: DEBUG nova.compute.utils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.128521] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 610.128714] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 610.264303] env[62552]: DEBUG nova.policy [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee6f13726f4a4cf9a9bf660606de018b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a52552bd7c184d3a97a1f27f658b0ea9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 610.312166] env[62552]: DEBUG nova.network.neutron [req-cd990b5a-12b5-48bd-b6fb-9e255d971fb2 req-09515146-2bb3-4b8a-abf5-d37152fe13dc service nova] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.359013] env[62552]: DEBUG nova.network.neutron [-] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.638162] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 610.644957] env[62552]: DEBUG nova.network.neutron [req-cd990b5a-12b5-48bd-b6fb-9e255d971fb2 req-09515146-2bb3-4b8a-abf5-d37152fe13dc service nova] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.687575] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79301db5-9b42-437a-a10d-a958b746bcfc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.695684] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a3b43f-4fa6-4a52-b4a4-6a609bdacab7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.732118] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4678cb-4818-439a-b3e4-e87de569147f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.742184] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0926931f-cecb-4e63-9487-c189307b3d6d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.757572] env[62552]: DEBUG nova.compute.provider_tree [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.862027] env[62552]: INFO nova.compute.manager [-] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Took 1.06 seconds to deallocate network for instance. [ 610.863947] env[62552]: DEBUG nova.compute.claims [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.864135] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.149048] env[62552]: DEBUG oslo_concurrency.lockutils [req-cd990b5a-12b5-48bd-b6fb-9e255d971fb2 req-09515146-2bb3-4b8a-abf5-d37152fe13dc service nova] Releasing lock "refresh_cache-92b9d5a7-e5ed-480e-b68e-b589e6d00832" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.149048] env[62552]: DEBUG nova.compute.manager [req-cd990b5a-12b5-48bd-b6fb-9e255d971fb2 req-09515146-2bb3-4b8a-abf5-d37152fe13dc service nova] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Received event network-vif-deleted-673852d0-7384-41cc-912f-b25d6edbb3d1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 611.263096] env[62552]: DEBUG nova.scheduler.client.report [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 611.317650] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Successfully created port: 70a0ceec-7c96-4714-8d80-d39c5ea54ade {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.652675] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 611.691783] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.691783] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.691957] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.692055] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.692485] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.692680] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.693522] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.693522] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.693522] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.693522] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.693669] env[62552]: DEBUG nova.virt.hardware [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.695749] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1f68b9-e5d3-41d2-a6cd-6e2e15305ca7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.706942] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9c9ef1-0a20-4164-bd7a-20e2072d6432 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.740227] env[62552]: ERROR nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 37c536bd-77ca-4c85-b9cb-65a824f863fe, please check neutron logs for more information. [ 611.740227] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.740227] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.740227] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.740227] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.740227] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.740227] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.740227] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.740227] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.740227] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 611.740227] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.740227] env[62552]: ERROR nova.compute.manager raise self.value [ 611.740227] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.740227] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.740227] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.740227] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.740840] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.740840] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.740840] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 37c536bd-77ca-4c85-b9cb-65a824f863fe, please check neutron logs for more information. [ 611.740840] env[62552]: ERROR nova.compute.manager [ 611.740840] env[62552]: Traceback (most recent call last): [ 611.740840] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.740840] env[62552]: listener.cb(fileno) [ 611.740840] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.740840] env[62552]: result = function(*args, **kwargs) [ 611.740840] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.740840] env[62552]: return func(*args, **kwargs) [ 611.740840] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 611.740840] env[62552]: raise e [ 611.740840] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.740840] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 611.740840] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.740840] env[62552]: created_port_ids = self._update_ports_for_instance( [ 611.740840] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.740840] env[62552]: with excutils.save_and_reraise_exception(): [ 611.740840] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.740840] env[62552]: self.force_reraise() [ 611.740840] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.740840] env[62552]: raise self.value [ 611.740840] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.740840] env[62552]: updated_port = self._update_port( [ 611.740840] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.740840] env[62552]: _ensure_no_port_binding_failure(port) [ 611.740840] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.740840] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.742081] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 37c536bd-77ca-4c85-b9cb-65a824f863fe, please check neutron logs for more information. [ 611.742081] env[62552]: Removing descriptor: 15 [ 611.742081] env[62552]: ERROR nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 37c536bd-77ca-4c85-b9cb-65a824f863fe, please check neutron logs for more information. [ 611.742081] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Traceback (most recent call last): [ 611.742081] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 611.742081] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] yield resources [ 611.742081] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 611.742081] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self.driver.spawn(context, instance, image_meta, [ 611.742081] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 611.742081] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.742081] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.742081] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] vm_ref = self.build_virtual_machine(instance, [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] for vif in network_info: [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] return self._sync_wrapper(fn, *args, **kwargs) [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self.wait() [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self[:] = self._gt.wait() [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] return self._exit_event.wait() [ 611.742499] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] result = hub.switch() [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] return self.greenlet.switch() [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] result = function(*args, **kwargs) [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] return func(*args, **kwargs) [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] raise e [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] nwinfo = self.network_api.allocate_for_instance( [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.742959] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] created_port_ids = self._update_ports_for_instance( [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] with excutils.save_and_reraise_exception(): [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self.force_reraise() [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] raise self.value [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] updated_port = self._update_port( [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] _ensure_no_port_binding_failure(port) [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.743464] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] raise exception.PortBindingFailed(port_id=port['id']) [ 611.743883] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] nova.exception.PortBindingFailed: Binding failed for port 37c536bd-77ca-4c85-b9cb-65a824f863fe, please check neutron logs for more information. [ 611.743883] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] [ 611.743883] env[62552]: INFO nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Terminating instance [ 611.768951] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.155s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.769736] env[62552]: ERROR nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4, please check neutron logs for more information. [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] Traceback (most recent call last): [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self.driver.spawn(context, instance, image_meta, [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] vm_ref = self.build_virtual_machine(instance, [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.769736] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] for vif in network_info: [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] return self._sync_wrapper(fn, *args, **kwargs) [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self.wait() [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self[:] = self._gt.wait() [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] return self._exit_event.wait() [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] result = hub.switch() [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.770628] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] return self.greenlet.switch() [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] result = function(*args, **kwargs) [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] return func(*args, **kwargs) [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] raise e [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] nwinfo = self.network_api.allocate_for_instance( [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] created_port_ids = self._update_ports_for_instance( [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] with excutils.save_and_reraise_exception(): [ 611.771197] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] self.force_reraise() [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] raise self.value [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] updated_port = self._update_port( [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] _ensure_no_port_binding_failure(port) [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] raise exception.PortBindingFailed(port_id=port['id']) [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] nova.exception.PortBindingFailed: Binding failed for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4, please check neutron logs for more information. [ 611.771678] env[62552]: ERROR nova.compute.manager [instance: fab97d93-2322-469d-84fa-812e1d491030] [ 611.772072] env[62552]: DEBUG nova.compute.utils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Binding failed for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.772785] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.504s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.774405] env[62552]: INFO nova.compute.claims [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.777735] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Build of instance fab97d93-2322-469d-84fa-812e1d491030 was re-scheduled: Binding failed for port 6e71a8c2-6b86-47c7-85cf-a31825c39bc4, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 611.778209] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 611.778451] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Acquiring lock "refresh_cache-fab97d93-2322-469d-84fa-812e1d491030" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.778595] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Acquired lock "refresh_cache-fab97d93-2322-469d-84fa-812e1d491030" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.779779] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.890654] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Acquiring lock "1c7e408a-392d-4fcf-bae8-588ed9f92fec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.890952] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Lock "1c7e408a-392d-4fcf-bae8-588ed9f92fec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.046525] env[62552]: DEBUG nova.compute.manager [req-e2119a46-7ee1-4802-acea-35e393395c40 req-76f0eb78-1b38-4d55-b22e-f05ded05416d service nova] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Received event network-changed-37c536bd-77ca-4c85-b9cb-65a824f863fe {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 612.047175] env[62552]: DEBUG nova.compute.manager [req-e2119a46-7ee1-4802-acea-35e393395c40 req-76f0eb78-1b38-4d55-b22e-f05ded05416d service nova] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Refreshing instance network info cache due to event network-changed-37c536bd-77ca-4c85-b9cb-65a824f863fe. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 612.047175] env[62552]: DEBUG oslo_concurrency.lockutils [req-e2119a46-7ee1-4802-acea-35e393395c40 req-76f0eb78-1b38-4d55-b22e-f05ded05416d service nova] Acquiring lock "refresh_cache-a34ea7fb-bd17-4353-806d-4cd2f5ee169d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.047175] env[62552]: DEBUG oslo_concurrency.lockutils [req-e2119a46-7ee1-4802-acea-35e393395c40 req-76f0eb78-1b38-4d55-b22e-f05ded05416d service nova] Acquired lock "refresh_cache-a34ea7fb-bd17-4353-806d-4cd2f5ee169d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.047175] env[62552]: DEBUG nova.network.neutron [req-e2119a46-7ee1-4802-acea-35e393395c40 req-76f0eb78-1b38-4d55-b22e-f05ded05416d service nova] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Refreshing network info cache for port 37c536bd-77ca-4c85-b9cb-65a824f863fe {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 612.246926] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Acquiring lock "refresh_cache-a34ea7fb-bd17-4353-806d-4cd2f5ee169d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.319243] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.572024] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.596533] env[62552]: DEBUG nova.network.neutron [req-e2119a46-7ee1-4802-acea-35e393395c40 req-76f0eb78-1b38-4d55-b22e-f05ded05416d service nova] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.934611] env[62552]: DEBUG nova.network.neutron [req-e2119a46-7ee1-4802-acea-35e393395c40 req-76f0eb78-1b38-4d55-b22e-f05ded05416d service nova] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.079618] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Releasing lock "refresh_cache-fab97d93-2322-469d-84fa-812e1d491030" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.079618] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 613.079618] env[62552]: DEBUG nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 613.082805] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.154852] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.428312] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751e406c-9771-4f9d-b17e-2dab396d17fd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.437417] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b20d7e6-4ef1-4854-b0ac-235adc762eaa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.444220] env[62552]: DEBUG oslo_concurrency.lockutils [req-e2119a46-7ee1-4802-acea-35e393395c40 req-76f0eb78-1b38-4d55-b22e-f05ded05416d service nova] Releasing lock "refresh_cache-a34ea7fb-bd17-4353-806d-4cd2f5ee169d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.444220] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Acquired lock "refresh_cache-a34ea7fb-bd17-4353-806d-4cd2f5ee169d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.444220] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.482024] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92bac0e-8e9a-4f8d-bb3f-ad57f83ea899 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.491089] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff501d34-a66d-406f-ac5d-fdf201d328d4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.508046] env[62552]: DEBUG nova.compute.provider_tree [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.658126] env[62552]: DEBUG nova.network.neutron [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.002406] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.014019] env[62552]: DEBUG nova.scheduler.client.report [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 614.163027] env[62552]: INFO nova.compute.manager [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] [instance: fab97d93-2322-469d-84fa-812e1d491030] Took 1.08 seconds to deallocate network for instance. [ 614.319794] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.523455] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.748s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.523455] env[62552]: DEBUG nova.compute.manager [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 614.525172] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.587s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.526890] env[62552]: INFO nova.compute.claims [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.719547] env[62552]: DEBUG nova.compute.manager [req-2d9354cd-20b1-4924-ba14-a033503cd1f6 req-d6173c89-eb6c-4938-91ac-9ce1e2498835 service nova] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Received event network-vif-deleted-37c536bd-77ca-4c85-b9cb-65a824f863fe {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 614.824132] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Releasing lock "refresh_cache-a34ea7fb-bd17-4353-806d-4cd2f5ee169d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.824573] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 614.824775] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 614.825424] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d05ba213-53ae-46b6-af4d-c39e543687c6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.834576] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a43193e-1de8-45cb-af12-706a86cc4fb2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.854752] env[62552]: ERROR nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade, please check neutron logs for more information. [ 614.854752] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.854752] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 614.854752] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.854752] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.854752] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.854752] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.854752] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.854752] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.854752] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 614.854752] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.854752] env[62552]: ERROR nova.compute.manager raise self.value [ 614.854752] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.854752] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.854752] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.854752] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.855401] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.855401] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.855401] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade, please check neutron logs for more information. [ 614.855401] env[62552]: ERROR nova.compute.manager [ 614.855401] env[62552]: Traceback (most recent call last): [ 614.855401] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.855401] env[62552]: listener.cb(fileno) [ 614.855401] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.855401] env[62552]: result = function(*args, **kwargs) [ 614.855401] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.855401] env[62552]: return func(*args, **kwargs) [ 614.855401] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 614.855401] env[62552]: raise e [ 614.855401] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 614.855401] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 614.855401] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.855401] env[62552]: created_port_ids = self._update_ports_for_instance( [ 614.855401] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.855401] env[62552]: with excutils.save_and_reraise_exception(): [ 614.855401] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.855401] env[62552]: self.force_reraise() [ 614.855401] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.855401] env[62552]: raise self.value [ 614.855401] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.855401] env[62552]: updated_port = self._update_port( [ 614.855401] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.855401] env[62552]: _ensure_no_port_binding_failure(port) [ 614.855401] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.855401] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.856396] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade, please check neutron logs for more information. [ 614.856396] env[62552]: Removing descriptor: 19 [ 614.856396] env[62552]: ERROR nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade, please check neutron logs for more information. [ 614.856396] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] Traceback (most recent call last): [ 614.856396] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 614.856396] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] yield resources [ 614.856396] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 614.856396] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self.driver.spawn(context, instance, image_meta, [ 614.856396] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 614.856396] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.856396] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.856396] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] vm_ref = self.build_virtual_machine(instance, [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] for vif in network_info: [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] return self._sync_wrapper(fn, *args, **kwargs) [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self.wait() [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self[:] = self._gt.wait() [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] return self._exit_event.wait() [ 614.856833] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] result = hub.switch() [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] return self.greenlet.switch() [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] result = function(*args, **kwargs) [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] return func(*args, **kwargs) [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] raise e [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] nwinfo = self.network_api.allocate_for_instance( [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.857285] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] created_port_ids = self._update_ports_for_instance( [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] with excutils.save_and_reraise_exception(): [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self.force_reraise() [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] raise self.value [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] updated_port = self._update_port( [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] _ensure_no_port_binding_failure(port) [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.857754] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] raise exception.PortBindingFailed(port_id=port['id']) [ 614.859455] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] nova.exception.PortBindingFailed: Binding failed for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade, please check neutron logs for more information. [ 614.859455] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] [ 614.859455] env[62552]: INFO nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Terminating instance [ 614.862718] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a34ea7fb-bd17-4353-806d-4cd2f5ee169d could not be found. [ 614.862926] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 614.863123] env[62552]: INFO nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 614.863368] env[62552]: DEBUG oslo.service.loopingcall [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.864559] env[62552]: DEBUG nova.compute.manager [-] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 614.864690] env[62552]: DEBUG nova.network.neutron [-] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.906362] env[62552]: DEBUG nova.network.neutron [-] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.034030] env[62552]: DEBUG nova.compute.utils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.038761] env[62552]: DEBUG nova.compute.manager [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Not allocating networking since 'none' was specified. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 615.201053] env[62552]: INFO nova.scheduler.client.report [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Deleted allocations for instance fab97d93-2322-469d-84fa-812e1d491030 [ 615.367692] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "refresh_cache-58c91f56-a683-4a62-bd45-80c627622621" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.369159] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquired lock "refresh_cache-58c91f56-a683-4a62-bd45-80c627622621" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.369159] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 615.409752] env[62552]: DEBUG nova.network.neutron [-] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.542413] env[62552]: DEBUG nova.compute.manager [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 615.712254] env[62552]: DEBUG oslo_concurrency.lockutils [None req-aad8b819-e8ed-42e5-859e-116782dd4302 tempest-ServerDiagnosticsTest-476555119 tempest-ServerDiagnosticsTest-476555119-project-member] Lock "fab97d93-2322-469d-84fa-812e1d491030" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.088s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.777086] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Acquiring lock "1252f45a-d55d-44c0-8629-aba51fb89956" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.778055] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Lock "1252f45a-d55d-44c0-8629-aba51fb89956" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.896537] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.911707] env[62552]: INFO nova.compute.manager [-] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Took 1.05 seconds to deallocate network for instance. [ 615.918010] env[62552]: DEBUG nova.compute.claims [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 615.918143] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.024839] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.194122] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ee15f8-b554-411b-aad5-2dc131ae95d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.202996] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33802e9-ffea-4217-b72b-46eb7b1e9a36 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.239083] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 616.242850] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7701bca0-a87a-4b82-bdf4-a7e45183718b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.250612] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edfadfd-3634-46da-bd56-7841e0cc2590 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.265298] env[62552]: DEBUG nova.compute.provider_tree [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.533880] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Releasing lock "refresh_cache-58c91f56-a683-4a62-bd45-80c627622621" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.534452] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 616.536751] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 616.536751] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5fe835b3-15bc-45e7-a791-2ddb25bfa4f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.546296] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a428233-2b15-4115-8fc5-b9aecf8d38e4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.560144] env[62552]: DEBUG nova.compute.manager [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 616.575031] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 58c91f56-a683-4a62-bd45-80c627622621 could not be found. [ 616.575031] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 616.575031] env[62552]: INFO nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Took 0.04 seconds to destroy the instance on the hypervisor. [ 616.576021] env[62552]: DEBUG oslo.service.loopingcall [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.576021] env[62552]: DEBUG nova.compute.manager [-] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 616.576021] env[62552]: DEBUG nova.network.neutron [-] [instance: 58c91f56-a683-4a62-bd45-80c627622621] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 616.585378] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.585620] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.585778] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.585960] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.586129] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.586281] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.586869] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.587162] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.587404] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.587607] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.587820] env[62552]: DEBUG nova.virt.hardware [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.588704] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413fd450-845e-46f2-bfc3-e01be7f3da57 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.596770] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9363ce67-a034-4a66-9e2a-4850ca0b242e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.611842] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 616.618914] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Creating folder: Project (1e50219e174c4bb4b3d415be27226406). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 616.619982] env[62552]: DEBUG nova.network.neutron [-] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.621268] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9daa74bf-2a01-4b56-9f02-f83c07666404 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.631547] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Created folder: Project (1e50219e174c4bb4b3d415be27226406) in parent group-v267339. [ 616.631547] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Creating folder: Instances. Parent ref: group-v267351. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 616.631547] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2e6c60ae-33f4-46ca-a552-bca500ecea1c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.640963] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Created folder: Instances in parent group-v267351. [ 616.641330] env[62552]: DEBUG oslo.service.loopingcall [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.641569] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 616.641778] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36165739-674c-4f75-bb2a-d45e82639f29 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.660568] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 616.660568] env[62552]: value = "task-1239217" [ 616.660568] env[62552]: _type = "Task" [ 616.660568] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.672077] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239217, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.770508] env[62552]: DEBUG nova.scheduler.client.report [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 616.775039] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.937169] env[62552]: DEBUG nova.compute.manager [req-fe69d604-9054-4201-ab95-677d98590283 req-05ca7e2d-1440-422d-85ac-e1cbad09e4f2 service nova] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Received event network-changed-70a0ceec-7c96-4714-8d80-d39c5ea54ade {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 616.937235] env[62552]: DEBUG nova.compute.manager [req-fe69d604-9054-4201-ab95-677d98590283 req-05ca7e2d-1440-422d-85ac-e1cbad09e4f2 service nova] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Refreshing instance network info cache due to event network-changed-70a0ceec-7c96-4714-8d80-d39c5ea54ade. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 616.937444] env[62552]: DEBUG oslo_concurrency.lockutils [req-fe69d604-9054-4201-ab95-677d98590283 req-05ca7e2d-1440-422d-85ac-e1cbad09e4f2 service nova] Acquiring lock "refresh_cache-58c91f56-a683-4a62-bd45-80c627622621" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.937577] env[62552]: DEBUG oslo_concurrency.lockutils [req-fe69d604-9054-4201-ab95-677d98590283 req-05ca7e2d-1440-422d-85ac-e1cbad09e4f2 service nova] Acquired lock "refresh_cache-58c91f56-a683-4a62-bd45-80c627622621" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.937733] env[62552]: DEBUG nova.network.neutron [req-fe69d604-9054-4201-ab95-677d98590283 req-05ca7e2d-1440-422d-85ac-e1cbad09e4f2 service nova] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Refreshing network info cache for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 617.057511] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "cdee22a9-4327-47af-ab14-8403a06b2802" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.057750] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.124762] env[62552]: DEBUG nova.network.neutron [-] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.171387] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239217, 'name': CreateVM_Task, 'duration_secs': 0.300126} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.171585] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 617.172060] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.172235] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.172635] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 617.172997] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0265b358-e3e9-4c04-8bc6-2963c8645fb4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.177240] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 617.177240] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a59201-4779-869f-a02b-a0b386a50759" [ 617.177240] env[62552]: _type = "Task" [ 617.177240] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.184426] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a59201-4779-869f-a02b-a0b386a50759, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.276916] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.752s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.277463] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 617.280142] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.390s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.281053] env[62552]: DEBUG nova.objects.instance [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62552) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 617.458748] env[62552]: DEBUG nova.network.neutron [req-fe69d604-9054-4201-ab95-677d98590283 req-05ca7e2d-1440-422d-85ac-e1cbad09e4f2 service nova] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.566843] env[62552]: DEBUG nova.network.neutron [req-fe69d604-9054-4201-ab95-677d98590283 req-05ca7e2d-1440-422d-85ac-e1cbad09e4f2 service nova] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.627939] env[62552]: INFO nova.compute.manager [-] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Took 1.05 seconds to deallocate network for instance. [ 617.630275] env[62552]: DEBUG nova.compute.claims [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 617.630449] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.688269] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a59201-4779-869f-a02b-a0b386a50759, 'name': SearchDatastore_Task, 'duration_secs': 0.009929} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.688647] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.688930] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 617.689415] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.689415] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.689615] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 617.689881] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3010a3ba-290c-41d9-9385-38355d2bc5e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.697661] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 617.697830] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 617.698600] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a14a32de-13f5-43d4-8b64-3d03ee06709c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.703835] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 617.703835] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e7fe87-015b-fb6a-cd75-f44c06c26342" [ 617.703835] env[62552]: _type = "Task" [ 617.703835] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.712482] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e7fe87-015b-fb6a-cd75-f44c06c26342, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.785169] env[62552]: DEBUG nova.compute.utils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.789210] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 617.789374] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.839579] env[62552]: DEBUG nova.policy [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df6790b1ecda468cb49fa26824c6d136', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd6a470be93504de3be6cd404cf860815', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 618.075721] env[62552]: DEBUG oslo_concurrency.lockutils [req-fe69d604-9054-4201-ab95-677d98590283 req-05ca7e2d-1440-422d-85ac-e1cbad09e4f2 service nova] Releasing lock "refresh_cache-58c91f56-a683-4a62-bd45-80c627622621" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.075721] env[62552]: DEBUG nova.compute.manager [req-fe69d604-9054-4201-ab95-677d98590283 req-05ca7e2d-1440-422d-85ac-e1cbad09e4f2 service nova] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Received event network-vif-deleted-70a0ceec-7c96-4714-8d80-d39c5ea54ade {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 618.214664] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e7fe87-015b-fb6a-cd75-f44c06c26342, 'name': SearchDatastore_Task, 'duration_secs': 0.009161} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.215492] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d052d40-de74-4c51-81f2-57044eaeb497 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.220754] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 618.220754] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52393fb4-5611-3f15-c4b3-12ebbaae5cfd" [ 618.220754] env[62552]: _type = "Task" [ 618.220754] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.231769] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52393fb4-5611-3f15-c4b3-12ebbaae5cfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.290613] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 618.295554] env[62552]: DEBUG oslo_concurrency.lockutils [None req-43117fd9-a25a-4bcc-b392-1d26ef01a775 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.296627] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.324s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.297958] env[62552]: INFO nova.compute.claims [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.390568] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Successfully created port: 47679719-5fb5-4c16-a0e2-ca9b2a678e71 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.739600] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52393fb4-5611-3f15-c4b3-12ebbaae5cfd, 'name': SearchDatastore_Task, 'duration_secs': 0.009033} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.739600] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.739906] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 618.739994] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9aab7302-b697-4489-89ac-9ab1ed57e9d7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.747973] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 618.747973] env[62552]: value = "task-1239222" [ 618.747973] env[62552]: _type = "Task" [ 618.747973] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.755820] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239222, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.258009] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239222, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449238} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.258293] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 619.258510] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 619.258774] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-817d181b-b2e5-444b-93ff-36c74727778f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.266207] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 619.266207] env[62552]: value = "task-1239224" [ 619.266207] env[62552]: _type = "Task" [ 619.266207] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.272542] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239224, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.307991] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 619.343659] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "1472ca01-ba04-4e7d-a9cf-ccaf29c04282" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.343809] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "1472ca01-ba04-4e7d-a9cf-ccaf29c04282" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.352833] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.352982] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.353109] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.353747] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.353747] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.353747] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.353747] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.354054] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.354253] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.354253] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.354985] env[62552]: DEBUG nova.virt.hardware [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.355212] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc90b730-889a-4cb7-ab4a-7741ce50b396 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.365087] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ad385e-2eaa-41f4-ad12-00dc50fe9596 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.659341] env[62552]: DEBUG nova.compute.manager [req-8a4801dc-8378-40a9-b7d1-dc68d930fe74 req-7a5bd9f7-e9ca-4066-ac41-4efd1bf05c90 service nova] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Received event network-changed-47679719-5fb5-4c16-a0e2-ca9b2a678e71 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 619.659622] env[62552]: DEBUG nova.compute.manager [req-8a4801dc-8378-40a9-b7d1-dc68d930fe74 req-7a5bd9f7-e9ca-4066-ac41-4efd1bf05c90 service nova] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Refreshing instance network info cache due to event network-changed-47679719-5fb5-4c16-a0e2-ca9b2a678e71. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 619.660691] env[62552]: DEBUG oslo_concurrency.lockutils [req-8a4801dc-8378-40a9-b7d1-dc68d930fe74 req-7a5bd9f7-e9ca-4066-ac41-4efd1bf05c90 service nova] Acquiring lock "refresh_cache-729eba7d-5f14-4506-8fee-24f6eb685b12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.660907] env[62552]: DEBUG oslo_concurrency.lockutils [req-8a4801dc-8378-40a9-b7d1-dc68d930fe74 req-7a5bd9f7-e9ca-4066-ac41-4efd1bf05c90 service nova] Acquired lock "refresh_cache-729eba7d-5f14-4506-8fee-24f6eb685b12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.662113] env[62552]: DEBUG nova.network.neutron [req-8a4801dc-8378-40a9-b7d1-dc68d930fe74 req-7a5bd9f7-e9ca-4066-ac41-4efd1bf05c90 service nova] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Refreshing network info cache for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 619.782751] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239224, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065874} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.783049] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 619.783818] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92aeec8e-1381-4bae-a5d8-188a926da8a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.805733] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 619.806616] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69513b5e-02d7-4c8c-81d3-322fccce606a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.829392] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 619.829392] env[62552]: value = "task-1239225" [ 619.829392] env[62552]: _type = "Task" [ 619.829392] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.837255] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239225, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.886944] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841bf61c-1af9-4410-8c12-9c3b73cf8fa2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.897321] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88105ff6-c421-4dcb-be50-0bcdc7380e82 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.930617] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d78a6a-2cd4-4629-84c4-ebed28f5d537 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.940502] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19103c5-d643-4a69-841b-11b70c718869 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.962166] env[62552]: DEBUG nova.compute.provider_tree [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.024899] env[62552]: ERROR nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71, please check neutron logs for more information. [ 620.024899] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 620.024899] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.024899] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 620.024899] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.024899] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 620.024899] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.024899] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 620.024899] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.024899] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 620.024899] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.024899] env[62552]: ERROR nova.compute.manager raise self.value [ 620.024899] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.024899] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 620.024899] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.024899] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 620.025473] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.025473] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 620.025473] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71, please check neutron logs for more information. [ 620.025473] env[62552]: ERROR nova.compute.manager [ 620.025473] env[62552]: Traceback (most recent call last): [ 620.025473] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 620.025473] env[62552]: listener.cb(fileno) [ 620.025473] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.025473] env[62552]: result = function(*args, **kwargs) [ 620.025473] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.025473] env[62552]: return func(*args, **kwargs) [ 620.025473] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 620.025473] env[62552]: raise e [ 620.025473] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.025473] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 620.025473] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.025473] env[62552]: created_port_ids = self._update_ports_for_instance( [ 620.025473] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.025473] env[62552]: with excutils.save_and_reraise_exception(): [ 620.025473] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.025473] env[62552]: self.force_reraise() [ 620.025473] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.025473] env[62552]: raise self.value [ 620.025473] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.025473] env[62552]: updated_port = self._update_port( [ 620.025473] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.025473] env[62552]: _ensure_no_port_binding_failure(port) [ 620.025473] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.025473] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 620.026423] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71, please check neutron logs for more information. [ 620.026423] env[62552]: Removing descriptor: 19 [ 620.026423] env[62552]: ERROR nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71, please check neutron logs for more information. [ 620.026423] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Traceback (most recent call last): [ 620.026423] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 620.026423] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] yield resources [ 620.026423] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 620.026423] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self.driver.spawn(context, instance, image_meta, [ 620.026423] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 620.026423] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.026423] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.026423] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] vm_ref = self.build_virtual_machine(instance, [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] for vif in network_info: [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] return self._sync_wrapper(fn, *args, **kwargs) [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self.wait() [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self[:] = self._gt.wait() [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] return self._exit_event.wait() [ 620.027017] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] result = hub.switch() [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] return self.greenlet.switch() [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] result = function(*args, **kwargs) [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] return func(*args, **kwargs) [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] raise e [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] nwinfo = self.network_api.allocate_for_instance( [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.027508] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] created_port_ids = self._update_ports_for_instance( [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] with excutils.save_and_reraise_exception(): [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self.force_reraise() [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] raise self.value [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] updated_port = self._update_port( [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] _ensure_no_port_binding_failure(port) [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.028091] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] raise exception.PortBindingFailed(port_id=port['id']) [ 620.028616] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] nova.exception.PortBindingFailed: Binding failed for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71, please check neutron logs for more information. [ 620.028616] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] [ 620.028616] env[62552]: INFO nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Terminating instance [ 620.183406] env[62552]: DEBUG nova.network.neutron [req-8a4801dc-8378-40a9-b7d1-dc68d930fe74 req-7a5bd9f7-e9ca-4066-ac41-4efd1bf05c90 service nova] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.287543] env[62552]: DEBUG nova.network.neutron [req-8a4801dc-8378-40a9-b7d1-dc68d930fe74 req-7a5bd9f7-e9ca-4066-ac41-4efd1bf05c90 service nova] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.342975] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239225, 'name': ReconfigVM_Task, 'duration_secs': 0.386081} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.343286] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 620.344523] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c17b64a0-48fc-4695-96e0-b311fac83347 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.353334] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 620.353334] env[62552]: value = "task-1239227" [ 620.353334] env[62552]: _type = "Task" [ 620.353334] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.360483] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239227, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.467817] env[62552]: DEBUG nova.scheduler.client.report [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 620.531828] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Acquiring lock "refresh_cache-729eba7d-5f14-4506-8fee-24f6eb685b12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.791722] env[62552]: DEBUG oslo_concurrency.lockutils [req-8a4801dc-8378-40a9-b7d1-dc68d930fe74 req-7a5bd9f7-e9ca-4066-ac41-4efd1bf05c90 service nova] Releasing lock "refresh_cache-729eba7d-5f14-4506-8fee-24f6eb685b12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.792127] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Acquired lock "refresh_cache-729eba7d-5f14-4506-8fee-24f6eb685b12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.792316] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 620.862902] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239227, 'name': Rename_Task, 'duration_secs': 0.303497} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.863190] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 620.863433] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-031d2f6b-ea82-47e5-854b-50a9fcb89245 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.869544] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 620.869544] env[62552]: value = "task-1239228" [ 620.869544] env[62552]: _type = "Task" [ 620.869544] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.877216] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239228, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.974739] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.677s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.975503] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 620.978945] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.121s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.979233] env[62552]: DEBUG nova.objects.instance [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lazy-loading 'resources' on Instance uuid 7408a538-6091-4aa2-b2d0-a3d93840b341 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 621.310791] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.379261] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239228, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.453069] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.481894] env[62552]: DEBUG nova.compute.utils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.483269] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 621.483437] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 621.556409] env[62552]: DEBUG nova.policy [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3df7e9a714dc4b988f731cbab5ce794d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e8b820039a1a4885b453de81bfbfc03d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 621.884793] env[62552]: DEBUG oslo_vmware.api [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239228, 'name': PowerOnVM_Task, 'duration_secs': 0.523096} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.887158] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 621.887158] env[62552]: INFO nova.compute.manager [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Took 5.33 seconds to spawn the instance on the hypervisor. [ 621.887158] env[62552]: DEBUG nova.compute.manager [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 621.887808] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8eefdbf-9ded-4206-a96f-80ad2030a342 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.949690] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3aa221d-cda2-4766-86dd-845e7d5d3fa4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.955389] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Releasing lock "refresh_cache-729eba7d-5f14-4506-8fee-24f6eb685b12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.957507] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 621.958465] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 621.958465] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-38bf26f9-fb97-48e6-8662-7b448d4cb6e4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.965810] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e392add2-1a7e-4da3-9991-158d549c1cb2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.973555] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977079b2-0149-458f-ae37-3baf47c224b3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.986402] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Successfully created port: 161e12f6-e3df-4714-baf4-ef7aaa99ed9c {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 621.990016] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 621.993176] env[62552]: DEBUG nova.compute.manager [req-e17f1370-2091-4a6b-b370-f29c63825fda req-e6a0a28a-4467-4092-9c58-e413093feb07 service nova] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Received event network-vif-deleted-47679719-5fb5-4c16-a0e2-ca9b2a678e71 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 622.025394] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa627c71-ba01-4188-b054-b414780ce2fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.036877] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf65416-4e3b-42da-a0da-b1af59c8acec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.041154] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 729eba7d-5f14-4506-8fee-24f6eb685b12 could not be found. [ 622.041370] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.041546] env[62552]: INFO nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Took 0.08 seconds to destroy the instance on the hypervisor. [ 622.042009] env[62552]: DEBUG oslo.service.loopingcall [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.042580] env[62552]: DEBUG nova.compute.manager [-] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 622.042683] env[62552]: DEBUG nova.network.neutron [-] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.054357] env[62552]: DEBUG nova.compute.provider_tree [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.069811] env[62552]: DEBUG nova.network.neutron [-] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.409989] env[62552]: INFO nova.compute.manager [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Took 39.16 seconds to build instance. [ 622.558238] env[62552]: DEBUG nova.scheduler.client.report [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 622.576528] env[62552]: DEBUG nova.network.neutron [-] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.631869] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Acquiring lock "8707ce4b-677e-4f13-86f8-3e327d19380b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.631869] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "8707ce4b-677e-4f13-86f8-3e327d19380b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.911707] env[62552]: DEBUG oslo_concurrency.lockutils [None req-163c4a3a-d8c8-4b27-ad0f-7cc6c3d2b6f1 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "9952d8f6-ee78-4c2d-b147-8c08c027f440" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.641s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.032341] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 623.054305] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.054594] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.054775] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.054964] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.055131] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.055278] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.055479] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.055699] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.055887] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.056058] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.056233] env[62552]: DEBUG nova.virt.hardware [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.057500] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ce905f-06ae-44b5-8756-20a314548741 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.065623] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fbb3a0-1904-4ada-8035-6ebd0b36ccd1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.071538] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.093s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.073654] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.224s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.076751] env[62552]: INFO nova.compute.manager [-] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Took 1.03 seconds to deallocate network for instance. [ 623.093034] env[62552]: DEBUG nova.compute.claims [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.093427] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.103822] env[62552]: INFO nova.scheduler.client.report [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Deleted allocations for instance 7408a538-6091-4aa2-b2d0-a3d93840b341 [ 623.414653] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 623.489047] env[62552]: INFO nova.compute.manager [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Rebuilding instance [ 623.537553] env[62552]: DEBUG nova.compute.manager [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 623.538459] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075e51db-0f26-42b9-aa03-96b3f4b834a1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.601648] env[62552]: ERROR nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c, please check neutron logs for more information. [ 623.601648] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 623.601648] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 623.601648] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 623.601648] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.601648] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 623.601648] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.601648] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 623.601648] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.601648] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 623.601648] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.601648] env[62552]: ERROR nova.compute.manager raise self.value [ 623.601648] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.601648] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 623.601648] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.601648] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 623.602019] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.602019] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 623.602019] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c, please check neutron logs for more information. [ 623.602019] env[62552]: ERROR nova.compute.manager [ 623.602019] env[62552]: Traceback (most recent call last): [ 623.602019] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 623.602019] env[62552]: listener.cb(fileno) [ 623.602019] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.602019] env[62552]: result = function(*args, **kwargs) [ 623.602019] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.602019] env[62552]: return func(*args, **kwargs) [ 623.602019] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 623.602019] env[62552]: raise e [ 623.602019] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 623.602019] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 623.602019] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.602019] env[62552]: created_port_ids = self._update_ports_for_instance( [ 623.602019] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.602019] env[62552]: with excutils.save_and_reraise_exception(): [ 623.602019] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.602019] env[62552]: self.force_reraise() [ 623.602019] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.602019] env[62552]: raise self.value [ 623.602019] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.602019] env[62552]: updated_port = self._update_port( [ 623.602019] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.602019] env[62552]: _ensure_no_port_binding_failure(port) [ 623.602019] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.602019] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 623.602558] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c, please check neutron logs for more information. [ 623.602558] env[62552]: Removing descriptor: 19 [ 623.602558] env[62552]: ERROR nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c, please check neutron logs for more information. [ 623.602558] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Traceback (most recent call last): [ 623.602558] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 623.602558] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] yield resources [ 623.602558] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 623.602558] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self.driver.spawn(context, instance, image_meta, [ 623.602558] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 623.602558] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.602558] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.602558] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] vm_ref = self.build_virtual_machine(instance, [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] for vif in network_info: [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] return self._sync_wrapper(fn, *args, **kwargs) [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self.wait() [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self[:] = self._gt.wait() [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] return self._exit_event.wait() [ 623.602877] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] result = hub.switch() [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] return self.greenlet.switch() [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] result = function(*args, **kwargs) [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] return func(*args, **kwargs) [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] raise e [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] nwinfo = self.network_api.allocate_for_instance( [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.603132] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] created_port_ids = self._update_ports_for_instance( [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] with excutils.save_and_reraise_exception(): [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self.force_reraise() [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] raise self.value [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] updated_port = self._update_port( [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] _ensure_no_port_binding_failure(port) [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.603432] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] raise exception.PortBindingFailed(port_id=port['id']) [ 623.603716] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] nova.exception.PortBindingFailed: Binding failed for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c, please check neutron logs for more information. [ 623.603716] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] [ 623.603716] env[62552]: INFO nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Terminating instance [ 623.616151] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81c39752-983f-4473-950a-87f6eb46a6d8 tempest-ServerShowV257Test-1288087721 tempest-ServerShowV257Test-1288087721-project-member] Lock "7408a538-6091-4aa2-b2d0-a3d93840b341" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.121s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.937165] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.006041] env[62552]: DEBUG nova.compute.manager [req-058215dc-055c-4ca2-b8db-d19b8767e1e7 req-5b67ff25-888c-483e-8836-eec2eaf27053 service nova] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Received event network-changed-161e12f6-e3df-4714-baf4-ef7aaa99ed9c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 624.006276] env[62552]: DEBUG nova.compute.manager [req-058215dc-055c-4ca2-b8db-d19b8767e1e7 req-5b67ff25-888c-483e-8836-eec2eaf27053 service nova] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Refreshing instance network info cache due to event network-changed-161e12f6-e3df-4714-baf4-ef7aaa99ed9c. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 624.006567] env[62552]: DEBUG oslo_concurrency.lockutils [req-058215dc-055c-4ca2-b8db-d19b8767e1e7 req-5b67ff25-888c-483e-8836-eec2eaf27053 service nova] Acquiring lock "refresh_cache-e22c0fef-3d0d-4339-be0b-f129b3628b80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.006744] env[62552]: DEBUG oslo_concurrency.lockutils [req-058215dc-055c-4ca2-b8db-d19b8767e1e7 req-5b67ff25-888c-483e-8836-eec2eaf27053 service nova] Acquired lock "refresh_cache-e22c0fef-3d0d-4339-be0b-f129b3628b80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.006937] env[62552]: DEBUG nova.network.neutron [req-058215dc-055c-4ca2-b8db-d19b8767e1e7 req-5b67ff25-888c-483e-8836-eec2eaf27053 service nova] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Refreshing network info cache for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 624.090688] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfcab1dc-2436-483a-80f4-0f0e871c59d4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.099999] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c83197-d8b6-4fb0-a9cb-268815757617 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.129190] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Acquiring lock "refresh_cache-e22c0fef-3d0d-4339-be0b-f129b3628b80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.131324] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4a6e4f-715b-46a2-b5cd-6dd8667bd2ac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.138095] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2f9278-9750-41a2-af42-32b3d6763fb0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.154938] env[62552]: DEBUG nova.compute.provider_tree [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.535246] env[62552]: DEBUG nova.network.neutron [req-058215dc-055c-4ca2-b8db-d19b8767e1e7 req-5b67ff25-888c-483e-8836-eec2eaf27053 service nova] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.555009] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 624.555502] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b8095c1-2b24-4cc7-ad3f-7de3b8d3e0c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.563587] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 624.563587] env[62552]: value = "task-1239231" [ 624.563587] env[62552]: _type = "Task" [ 624.563587] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.571589] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239231, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.658937] env[62552]: DEBUG nova.scheduler.client.report [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 624.671642] env[62552]: DEBUG nova.network.neutron [req-058215dc-055c-4ca2-b8db-d19b8767e1e7 req-5b67ff25-888c-483e-8836-eec2eaf27053 service nova] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.072912] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239231, 'name': PowerOffVM_Task, 'duration_secs': 0.210334} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.073205] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 625.073422] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.074175] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bcb3413-9e09-40e5-96a4-d0dd4cdf2b5a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.080424] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 625.080642] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8e345b7-27fb-4714-8ad6-e8a68940ef54 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.106758] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 625.106989] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 625.107184] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Deleting the datastore file [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 625.107435] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e1ee853-95dc-43da-9a7d-482a68604958 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.112837] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 625.112837] env[62552]: value = "task-1239233" [ 625.112837] env[62552]: _type = "Task" [ 625.112837] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.120032] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239233, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.165061] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.091s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.165720] env[62552]: ERROR nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5, please check neutron logs for more information. [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Traceback (most recent call last): [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self.driver.spawn(context, instance, image_meta, [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] vm_ref = self.build_virtual_machine(instance, [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.165720] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] for vif in network_info: [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] return self._sync_wrapper(fn, *args, **kwargs) [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self.wait() [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self[:] = self._gt.wait() [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] return self._exit_event.wait() [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] result = hub.switch() [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.166018] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] return self.greenlet.switch() [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] result = function(*args, **kwargs) [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] return func(*args, **kwargs) [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] raise e [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] nwinfo = self.network_api.allocate_for_instance( [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] created_port_ids = self._update_ports_for_instance( [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] with excutils.save_and_reraise_exception(): [ 625.166342] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] self.force_reraise() [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] raise self.value [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] updated_port = self._update_port( [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] _ensure_no_port_binding_failure(port) [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] raise exception.PortBindingFailed(port_id=port['id']) [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] nova.exception.PortBindingFailed: Binding failed for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5, please check neutron logs for more information. [ 625.166656] env[62552]: ERROR nova.compute.manager [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] [ 625.166939] env[62552]: DEBUG nova.compute.utils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Binding failed for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.167853] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.422s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.170875] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Build of instance c4a725a7-a14c-44ec-a81a-3e965d7e0817 was re-scheduled: Binding failed for port e5b1d0f8-689f-4a7f-99a1-52883e53bfa5, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 625.171319] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 625.171540] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "refresh_cache-c4a725a7-a14c-44ec-a81a-3e965d7e0817" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.171681] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquired lock "refresh_cache-c4a725a7-a14c-44ec-a81a-3e965d7e0817" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.171836] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.176852] env[62552]: DEBUG oslo_concurrency.lockutils [req-058215dc-055c-4ca2-b8db-d19b8767e1e7 req-5b67ff25-888c-483e-8836-eec2eaf27053 service nova] Releasing lock "refresh_cache-e22c0fef-3d0d-4339-be0b-f129b3628b80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.177095] env[62552]: DEBUG nova.compute.manager [req-058215dc-055c-4ca2-b8db-d19b8767e1e7 req-5b67ff25-888c-483e-8836-eec2eaf27053 service nova] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Received event network-vif-deleted-161e12f6-e3df-4714-baf4-ef7aaa99ed9c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 625.177456] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Acquired lock "refresh_cache-e22c0fef-3d0d-4339-be0b-f129b3628b80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.177629] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.624366] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239233, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.083917} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.625417] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 625.625417] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 625.625417] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.709386] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.712226] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.918246] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.922780] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.133641] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85d552f-2749-4a91-9905-6751e269464c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.141832] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b8519e-599f-4ad4-86e3-af516d26fbfe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.173411] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c73060-54c9-4348-b562-efa16bdc080b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.181185] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bbf63b-d1b6-4510-9738-0a344ce1857b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.195487] env[62552]: DEBUG nova.compute.provider_tree [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.423334] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Releasing lock "refresh_cache-c4a725a7-a14c-44ec-a81a-3e965d7e0817" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.423620] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 626.423816] env[62552]: DEBUG nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 626.424016] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.426494] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Releasing lock "refresh_cache-e22c0fef-3d0d-4339-be0b-f129b3628b80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.426898] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 626.427106] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.427411] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9d0fd85-c54f-49a1-ba95-457dfd6c6526 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.437530] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9c95e8-2e1f-4a7c-907d-41c7420e4245 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.449229] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.462555] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e22c0fef-3d0d-4339-be0b-f129b3628b80 could not be found. [ 626.462795] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 626.463110] env[62552]: INFO nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Took 0.04 seconds to destroy the instance on the hypervisor. [ 626.463323] env[62552]: DEBUG oslo.service.loopingcall [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.469051] env[62552]: DEBUG nova.compute.manager [-] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 626.469051] env[62552]: DEBUG nova.network.neutron [-] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.487845] env[62552]: DEBUG nova.network.neutron [-] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.663355] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.664745] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.664745] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.664745] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.664745] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.664745] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.665078] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.666181] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.666181] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.666181] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.666181] env[62552]: DEBUG nova.virt.hardware [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.667364] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9614fe1e-6005-4e06-9e92-d4bf1a581577 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.678030] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a677c1-e2db-475d-8d75-6cdb27f01de8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.692486] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 626.699070] env[62552]: DEBUG oslo.service.loopingcall [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.703021] env[62552]: DEBUG nova.scheduler.client.report [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 626.705067] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 626.705366] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44266e0b-8f86-4e0c-95e0-96874fc6378d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.726220] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 626.726220] env[62552]: value = "task-1239235" [ 626.726220] env[62552]: _type = "Task" [ 626.726220] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.734150] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239235, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.954527] env[62552]: DEBUG nova.network.neutron [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.991685] env[62552]: DEBUG nova.network.neutron [-] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.222535] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.054s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.222820] env[62552]: ERROR nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 74e5cee9-1240-45ee-84b4-f9afe2710f71, please check neutron logs for more information. [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Traceback (most recent call last): [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self.driver.spawn(context, instance, image_meta, [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] vm_ref = self.build_virtual_machine(instance, [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.222820] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] for vif in network_info: [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] return self._sync_wrapper(fn, *args, **kwargs) [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self.wait() [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self[:] = self._gt.wait() [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] return self._exit_event.wait() [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] result = hub.switch() [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.223123] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] return self.greenlet.switch() [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] result = function(*args, **kwargs) [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] return func(*args, **kwargs) [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] raise e [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] nwinfo = self.network_api.allocate_for_instance( [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] created_port_ids = self._update_ports_for_instance( [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] with excutils.save_and_reraise_exception(): [ 627.223442] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] self.force_reraise() [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] raise self.value [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] updated_port = self._update_port( [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] _ensure_no_port_binding_failure(port) [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] raise exception.PortBindingFailed(port_id=port['id']) [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] nova.exception.PortBindingFailed: Binding failed for port 74e5cee9-1240-45ee-84b4-f9afe2710f71, please check neutron logs for more information. [ 627.223757] env[62552]: ERROR nova.compute.manager [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] [ 627.224023] env[62552]: DEBUG nova.compute.utils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Binding failed for port 74e5cee9-1240-45ee-84b4-f9afe2710f71, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.224533] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.036s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.227976] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Build of instance 27e4d3de-3e44-4293-9b04-e511c9f02b42 was re-scheduled: Binding failed for port 74e5cee9-1240-45ee-84b4-f9afe2710f71, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 627.228529] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 627.229196] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "refresh_cache-27e4d3de-3e44-4293-9b04-e511c9f02b42" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.229196] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquired lock "refresh_cache-27e4d3de-3e44-4293-9b04-e511c9f02b42" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.229196] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.238475] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239235, 'name': CreateVM_Task, 'duration_secs': 0.31415} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.239154] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 627.239536] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.239711] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.240036] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 627.240497] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5524f205-a0da-4593-9f46-470199e5630e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.246110] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 627.246110] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5254744b-8aa7-424c-6c2c-0597733f3285" [ 627.246110] env[62552]: _type = "Task" [ 627.246110] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.253206] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5254744b-8aa7-424c-6c2c-0597733f3285, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.458031] env[62552]: INFO nova.compute.manager [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: c4a725a7-a14c-44ec-a81a-3e965d7e0817] Took 1.03 seconds to deallocate network for instance. [ 627.496358] env[62552]: INFO nova.compute.manager [-] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Took 1.03 seconds to deallocate network for instance. [ 627.498609] env[62552]: DEBUG nova.compute.claims [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.498830] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.757258] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5254744b-8aa7-424c-6c2c-0597733f3285, 'name': SearchDatastore_Task, 'duration_secs': 0.008671} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.757258] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.757258] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 627.757461] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.757461] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.757608] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 627.757872] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2ef9106-89d5-40f9-bd5d-f4ee416597b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.760647] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.766321] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 627.766840] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 627.767399] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84c48aeb-3dd5-4cc4-b0e4-44c6756e7665 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.772317] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 627.772317] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d9f590-a5dd-da63-6e36-d2d2f43346b5" [ 627.772317] env[62552]: _type = "Task" [ 627.772317] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.779738] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d9f590-a5dd-da63-6e36-d2d2f43346b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.876966] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.185122] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0ca544-4fab-4a31-ba5f-2cf3f72e173a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.192743] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9045c0b9-f836-4445-9da0-9130053f6327 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.224245] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ff4028-d1d7-4ff7-9a23-23068fe451ee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.232292] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66caedec-2b16-4c18-8a22-a2ef010681c6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.245741] env[62552]: DEBUG nova.compute.provider_tree [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.281990] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d9f590-a5dd-da63-6e36-d2d2f43346b5, 'name': SearchDatastore_Task, 'duration_secs': 0.01079} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.282795] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3acb6252-cbd7-489e-a576-52d9262d3c7b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.287800] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 628.287800] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5206a8ab-4eb9-a82c-7dda-925e2b0a1226" [ 628.287800] env[62552]: _type = "Task" [ 628.287800] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.295261] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5206a8ab-4eb9-a82c-7dda-925e2b0a1226, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.379713] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Releasing lock "refresh_cache-27e4d3de-3e44-4293-9b04-e511c9f02b42" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.380052] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 628.380244] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 628.380409] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.399405] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.486521] env[62552]: INFO nova.scheduler.client.report [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Deleted allocations for instance c4a725a7-a14c-44ec-a81a-3e965d7e0817 [ 628.748744] env[62552]: DEBUG nova.scheduler.client.report [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 628.798457] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5206a8ab-4eb9-a82c-7dda-925e2b0a1226, 'name': SearchDatastore_Task, 'duration_secs': 0.008357} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.798786] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.800028] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 628.800028] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0923d4ea-e892-454b-93e1-880833c13ec7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.806025] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 628.806025] env[62552]: value = "task-1239237" [ 628.806025] env[62552]: _type = "Task" [ 628.806025] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.813987] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.901773] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.994980] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c7c95cfe-e54c-4641-a272-20b9eb42433f tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "c4a725a7-a14c-44ec-a81a-3e965d7e0817" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.223s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.254196] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.030s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.254912] env[62552]: ERROR nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] Traceback (most recent call last): [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self.driver.spawn(context, instance, image_meta, [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] vm_ref = self.build_virtual_machine(instance, [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.254912] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] for vif in network_info: [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] return self._sync_wrapper(fn, *args, **kwargs) [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self.wait() [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self[:] = self._gt.wait() [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] return self._exit_event.wait() [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] result = hub.switch() [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.255208] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] return self.greenlet.switch() [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] result = function(*args, **kwargs) [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] return func(*args, **kwargs) [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] raise e [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] nwinfo = self.network_api.allocate_for_instance( [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] created_port_ids = self._update_ports_for_instance( [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] with excutils.save_and_reraise_exception(): [ 629.255472] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] self.force_reraise() [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] raise self.value [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] updated_port = self._update_port( [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] _ensure_no_port_binding_failure(port) [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] raise exception.PortBindingFailed(port_id=port['id']) [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] nova.exception.PortBindingFailed: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. [ 629.255743] env[62552]: ERROR nova.compute.manager [instance: a44de380-895c-4f06-835e-595917a25be1] [ 629.255983] env[62552]: DEBUG nova.compute.utils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.257041] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.393s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.261153] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Build of instance a44de380-895c-4f06-835e-595917a25be1 was re-scheduled: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 629.261153] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 629.261434] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "refresh_cache-a44de380-895c-4f06-835e-595917a25be1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.261612] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquired lock "refresh_cache-a44de380-895c-4f06-835e-595917a25be1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.261806] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.317323] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239237, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432945} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.317547] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 629.317756] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 629.317991] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03cd0aa1-19a1-45cd-b486-d77339d2cd6f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.324050] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 629.324050] env[62552]: value = "task-1239238" [ 629.324050] env[62552]: _type = "Task" [ 629.324050] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.331394] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239238, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.404019] env[62552]: INFO nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 27e4d3de-3e44-4293-9b04-e511c9f02b42] Took 1.02 seconds to deallocate network for instance. [ 629.498531] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 629.787125] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.836799] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239238, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123447} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.839445] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 629.840485] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d155ace6-6cfb-40c1-8b72-a26adee18550 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.860710] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 629.865293] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18ec2bdf-12e1-46d0-bbbb-ab1817228cde {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.884740] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 629.884740] env[62552]: value = "task-1239239" [ 629.884740] env[62552]: _type = "Task" [ 629.884740] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.893300] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.903140] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.021428] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.249517] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a37e5bd-7307-42b8-b9f6-e9518b1735bb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.257049] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b10ffb-ef7f-4860-96dc-66319850e81d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.288175] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a5e949-ab8c-44fd-b652-d15fa1bf131e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.294399] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469de81b-2bd7-47fc-9eea-b8772ad75ae3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.306976] env[62552]: DEBUG nova.compute.provider_tree [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.395807] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239239, 'name': ReconfigVM_Task, 'duration_secs': 0.27188} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.396496] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 630.397149] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-534bedf8-a14a-4fcf-8ddb-415002545ba1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.403400] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 630.403400] env[62552]: value = "task-1239241" [ 630.403400] env[62552]: _type = "Task" [ 630.403400] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.407439] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Releasing lock "refresh_cache-a44de380-895c-4f06-835e-595917a25be1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.407650] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 630.407830] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 630.407995] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.417445] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239241, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.430767] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.440648] env[62552]: INFO nova.scheduler.client.report [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Deleted allocations for instance 27e4d3de-3e44-4293-9b04-e511c9f02b42 [ 630.810615] env[62552]: DEBUG nova.scheduler.client.report [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 630.913424] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239241, 'name': Rename_Task, 'duration_secs': 0.159475} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.913822] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 630.914136] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-544d33f7-42b8-46cb-af5e-a5e56931de5e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.919803] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 630.919803] env[62552]: value = "task-1239242" [ 630.919803] env[62552]: _type = "Task" [ 630.919803] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.927465] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.933933] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.949133] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "27e4d3de-3e44-4293-9b04-e511c9f02b42" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.313s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.315831] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.059s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.316752] env[62552]: ERROR nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Traceback (most recent call last): [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self.driver.spawn(context, instance, image_meta, [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] vm_ref = self.build_virtual_machine(instance, [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.316752] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] for vif in network_info: [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] return self._sync_wrapper(fn, *args, **kwargs) [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self.wait() [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self[:] = self._gt.wait() [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] return self._exit_event.wait() [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] result = hub.switch() [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.317121] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] return self.greenlet.switch() [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] result = function(*args, **kwargs) [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] return func(*args, **kwargs) [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] raise e [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] nwinfo = self.network_api.allocate_for_instance( [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] created_port_ids = self._update_ports_for_instance( [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] with excutils.save_and_reraise_exception(): [ 631.317442] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] self.force_reraise() [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] raise self.value [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] updated_port = self._update_port( [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] _ensure_no_port_binding_failure(port) [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] raise exception.PortBindingFailed(port_id=port['id']) [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] nova.exception.PortBindingFailed: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. [ 631.317770] env[62552]: ERROR nova.compute.manager [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] [ 631.318085] env[62552]: DEBUG nova.compute.utils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.318290] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.400s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.321251] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Build of instance 92b9d5a7-e5ed-480e-b68e-b589e6d00832 was re-scheduled: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 631.321699] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 631.322392] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquiring lock "refresh_cache-92b9d5a7-e5ed-480e-b68e-b589e6d00832" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.322392] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Acquired lock "refresh_cache-92b9d5a7-e5ed-480e-b68e-b589e6d00832" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.322392] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.434237] env[62552]: DEBUG oslo_vmware.api [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239242, 'name': PowerOnVM_Task, 'duration_secs': 0.422505} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.434237] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 631.434237] env[62552]: DEBUG nova.compute.manager [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 631.434237] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8127e4a4-170e-4721-931c-2310012b3278 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.437309] env[62552]: INFO nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: a44de380-895c-4f06-835e-595917a25be1] Took 1.03 seconds to deallocate network for instance. [ 631.454165] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 631.851490] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.958629] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.972749] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.979212] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.265476] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa570399-8120-4c47-8ed0-a6b88b6b3ea7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.278041] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29ceeb9-1457-4ca7-b04e-d9235345d0df {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.308969] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178839f3-be84-4ddf-895f-5c438eb5610f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.316565] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7917c388-ec45-4bb3-af71-4d2e3f94a39a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.331954] env[62552]: DEBUG nova.compute.provider_tree [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.477823] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Releasing lock "refresh_cache-92b9d5a7-e5ed-480e-b68e-b589e6d00832" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.478106] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 632.478253] env[62552]: DEBUG nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 632.478434] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.518845] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.610391] env[62552]: INFO nova.compute.manager [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Rebuilding instance [ 632.662022] env[62552]: DEBUG nova.compute.manager [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 632.662022] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ced03e-371d-4a2c-8612-3e4ba2323fd3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.837561] env[62552]: DEBUG nova.scheduler.client.report [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 633.024204] env[62552]: DEBUG nova.network.neutron [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.343606] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.344270] env[62552]: ERROR nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 37c536bd-77ca-4c85-b9cb-65a824f863fe, please check neutron logs for more information. [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Traceback (most recent call last): [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self.driver.spawn(context, instance, image_meta, [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] vm_ref = self.build_virtual_machine(instance, [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.344270] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] for vif in network_info: [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] return self._sync_wrapper(fn, *args, **kwargs) [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self.wait() [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self[:] = self._gt.wait() [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] return self._exit_event.wait() [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] result = hub.switch() [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.344797] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] return self.greenlet.switch() [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] result = function(*args, **kwargs) [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] return func(*args, **kwargs) [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] raise e [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] nwinfo = self.network_api.allocate_for_instance( [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] created_port_ids = self._update_ports_for_instance( [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] with excutils.save_and_reraise_exception(): [ 633.345351] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] self.force_reraise() [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] raise self.value [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] updated_port = self._update_port( [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] _ensure_no_port_binding_failure(port) [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] raise exception.PortBindingFailed(port_id=port['id']) [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] nova.exception.PortBindingFailed: Binding failed for port 37c536bd-77ca-4c85-b9cb-65a824f863fe, please check neutron logs for more information. [ 633.345877] env[62552]: ERROR nova.compute.manager [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] [ 633.347287] env[62552]: DEBUG nova.compute.utils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Binding failed for port 37c536bd-77ca-4c85-b9cb-65a824f863fe, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.347287] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.571s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.348025] env[62552]: INFO nova.compute.claims [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.350508] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Build of instance a34ea7fb-bd17-4353-806d-4cd2f5ee169d was re-scheduled: Binding failed for port 37c536bd-77ca-4c85-b9cb-65a824f863fe, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 633.350942] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 633.351180] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Acquiring lock "refresh_cache-a34ea7fb-bd17-4353-806d-4cd2f5ee169d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.351325] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Acquired lock "refresh_cache-a34ea7fb-bd17-4353-806d-4cd2f5ee169d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.351483] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.486146] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "a44de380-895c-4f06-835e-595917a25be1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.874s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.486809] env[62552]: Traceback (most recent call last): [ 633.486855] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 633.486855] env[62552]: self.driver.spawn(context, instance, image_meta, [ 633.486855] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 633.486855] env[62552]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.486855] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.486855] env[62552]: vm_ref = self.build_virtual_machine(instance, [ 633.486999] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.486999] env[62552]: vif_infos = vmwarevif.get_vif_info(self._session, [ 633.486999] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.486999] env[62552]: for vif in network_info: [ 633.486999] env[62552]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.486999] env[62552]: return self._sync_wrapper(fn, *args, **kwargs) [ 633.486999] env[62552]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.486999] env[62552]: self.wait() [ 633.486999] env[62552]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.486999] env[62552]: self[:] = self._gt.wait() [ 633.486999] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.486999] env[62552]: return self._exit_event.wait() [ 633.486999] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.486999] env[62552]: result = hub.switch() [ 633.486999] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.486999] env[62552]: return self.greenlet.switch() [ 633.486999] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.486999] env[62552]: result = function(*args, **kwargs) [ 633.486999] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.486999] env[62552]: return func(*args, **kwargs) [ 633.486999] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 633.486999] env[62552]: raise e [ 633.486999] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 633.486999] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 633.486999] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.486999] env[62552]: created_port_ids = self._update_ports_for_instance( [ 633.486999] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.486999] env[62552]: with excutils.save_and_reraise_exception(): [ 633.486999] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.486999] env[62552]: self.force_reraise() [ 633.486999] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.486999] env[62552]: raise self.value [ 633.487769] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.487769] env[62552]: updated_port = self._update_port( [ 633.487769] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.487769] env[62552]: _ensure_no_port_binding_failure(port) [ 633.487769] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.487769] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.487769] env[62552]: nova.exception.PortBindingFailed: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. [ 633.487769] env[62552]: During handling of the above exception, another exception occurred: [ 633.487769] env[62552]: Traceback (most recent call last): [ 633.487769] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2462, in _do_build_and_run_instance [ 633.487769] env[62552]: self._build_and_run_instance(context, instance, image, [ 633.487769] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2754, in _build_and_run_instance [ 633.487769] env[62552]: raise exception.RescheduledException( [ 633.487769] env[62552]: nova.exception.RescheduledException: Build of instance a44de380-895c-4f06-835e-595917a25be1 was re-scheduled: Binding failed for port c059d502-667b-4152-bc99-bf65545180f7, please check neutron logs for more information. [ 633.487769] env[62552]: During handling of the above exception, another exception occurred: [ 633.487769] env[62552]: Traceback (most recent call last): [ 633.487769] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 633.487769] env[62552]: func(*args, **kwargs) [ 633.487769] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.487769] env[62552]: return func(*args, **kwargs) [ 633.487769] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 633.487769] env[62552]: return f(*args, **kwargs) [ 633.487769] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2353, in _locked_do_build_and_run_instance [ 633.487769] env[62552]: result = self._do_build_and_run_instance(*args, **kwargs) [ 633.487769] env[62552]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 633.487769] env[62552]: with excutils.save_and_reraise_exception(): [ 633.487769] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.487769] env[62552]: self.force_reraise() [ 633.488455] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.488455] env[62552]: raise self.value [ 633.488455] env[62552]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 633.488455] env[62552]: return f(self, context, *args, **kw) [ 633.488455] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 168, in decorated_function [ 633.488455] env[62552]: with excutils.save_and_reraise_exception(): [ 633.488455] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.488455] env[62552]: self.force_reraise() [ 633.488455] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.488455] env[62552]: raise self.value [ 633.488455] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 159, in decorated_function [ 633.488455] env[62552]: return function(self, context, *args, **kwargs) [ 633.488455] env[62552]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 633.488455] env[62552]: return function(self, context, *args, **kwargs) [ 633.488455] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 205, in decorated_function [ 633.488455] env[62552]: return function(self, context, *args, **kwargs) [ 633.488455] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2497, in _do_build_and_run_instance [ 633.488455] env[62552]: instance.save() [ 633.488455] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 633.488455] env[62552]: updates, result = self.indirection_api.object_action( [ 633.488455] env[62552]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 633.488455] env[62552]: return cctxt.call(context, 'object_action', objinst=objinst, [ 633.488455] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 633.488455] env[62552]: result = self.transport._send( [ 633.488455] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 633.488455] env[62552]: return self._driver.send(target, ctxt, message, [ 633.488455] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 633.488455] env[62552]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 633.489144] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 633.489144] env[62552]: raise result [ 633.489144] env[62552]: nova.exception_Remote.InstanceNotFound_Remote: Instance a44de380-895c-4f06-835e-595917a25be1 could not be found. [ 633.489144] env[62552]: Traceback (most recent call last): [ 633.489144] env[62552]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 633.489144] env[62552]: return getattr(target, method)(*args, **kwargs) [ 633.489144] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 633.489144] env[62552]: return fn(self, *args, **kwargs) [ 633.489144] env[62552]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 633.489144] env[62552]: old_ref, inst_ref = db.instance_update_and_get_original( [ 633.489144] env[62552]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 633.489144] env[62552]: return f(*args, **kwargs) [ 633.489144] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 633.489144] env[62552]: with excutils.save_and_reraise_exception() as ectxt: [ 633.489144] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.489144] env[62552]: self.force_reraise() [ 633.489144] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.489144] env[62552]: raise self.value [ 633.489144] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 633.489144] env[62552]: return f(*args, **kwargs) [ 633.489144] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 633.489144] env[62552]: return f(context, *args, **kwargs) [ 633.489144] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 633.489144] env[62552]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 633.489144] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 633.489144] env[62552]: raise exception.InstanceNotFound(instance_id=uuid) [ 633.489144] env[62552]: nova.exception.InstanceNotFound: Instance a44de380-895c-4f06-835e-595917a25be1 could not be found. [ 633.527429] env[62552]: INFO nova.compute.manager [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] [instance: 92b9d5a7-e5ed-480e-b68e-b589e6d00832] Took 1.05 seconds to deallocate network for instance. [ 633.677221] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 633.678123] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a24de9d2-cc0f-4d1f-9dab-65c69f2c661c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.686585] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 633.686585] env[62552]: value = "task-1239244" [ 633.686585] env[62552]: _type = "Task" [ 633.686585] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.695087] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239244, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.883265] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.993748] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 634.021302] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.199821] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239244, 'name': PowerOffVM_Task, 'duration_secs': 0.132914} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.199866] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 634.200153] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.201149] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84cc954d-825d-4695-9f44-3a8fbb8b450f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.207380] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 634.207929] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b1736c2-b9f5-4432-a03f-610ab9cbbb9a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.234810] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 634.235120] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 634.235311] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Deleting the datastore file [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 634.235578] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e59a2768-47f7-41af-9a79-da66d65ac7cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.244780] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 634.244780] env[62552]: value = "task-1239246" [ 634.244780] env[62552]: _type = "Task" [ 634.244780] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.252915] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239246, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.512125] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.523796] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Releasing lock "refresh_cache-a34ea7fb-bd17-4353-806d-4cd2f5ee169d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.524029] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 634.524213] env[62552]: DEBUG nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 634.524373] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.542095] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.755258] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239246, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087476} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.755258] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 634.755258] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 634.756111] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.773557] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514f5916-c44c-4f6b-8e58-fff81da1e557 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.783981] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4310933f-84dd-4aa8-9586-d80fbba5e76a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.818629] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492549a4-2323-43f1-8fed-6712386c2a82 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.826481] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db875ef-4a6e-49fa-8484-39e9bb72b04f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.840394] env[62552]: DEBUG nova.compute.provider_tree [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.046208] env[62552]: DEBUG nova.network.neutron [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.344101] env[62552]: DEBUG nova.scheduler.client.report [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 635.555898] env[62552]: INFO nova.compute.manager [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] [instance: a34ea7fb-bd17-4353-806d-4cd2f5ee169d] Took 1.03 seconds to deallocate network for instance. [ 635.569394] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0daf5afc-79e8-45c3-b1c0-789d506459cf tempest-ListServersNegativeTestJSON-1805803156 tempest-ListServersNegativeTestJSON-1805803156-project-member] Lock "92b9d5a7-e5ed-480e-b68e-b589e6d00832" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.865s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.569793] env[62552]: Traceback (most recent call last): [ 635.569846] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 635.569846] env[62552]: self.driver.spawn(context, instance, image_meta, [ 635.569846] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 635.569846] env[62552]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.569846] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.569846] env[62552]: vm_ref = self.build_virtual_machine(instance, [ 635.569846] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.569846] env[62552]: vif_infos = vmwarevif.get_vif_info(self._session, [ 635.570058] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.570058] env[62552]: for vif in network_info: [ 635.570058] env[62552]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.570058] env[62552]: return self._sync_wrapper(fn, *args, **kwargs) [ 635.570058] env[62552]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.570058] env[62552]: self.wait() [ 635.570058] env[62552]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.570058] env[62552]: self[:] = self._gt.wait() [ 635.570058] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.570058] env[62552]: return self._exit_event.wait() [ 635.570058] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.570058] env[62552]: result = hub.switch() [ 635.570058] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.570058] env[62552]: return self.greenlet.switch() [ 635.570058] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.570058] env[62552]: result = function(*args, **kwargs) [ 635.570058] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.570058] env[62552]: return func(*args, **kwargs) [ 635.570058] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 635.570058] env[62552]: raise e [ 635.570058] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.570058] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 635.570058] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.570058] env[62552]: created_port_ids = self._update_ports_for_instance( [ 635.570058] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.570058] env[62552]: with excutils.save_and_reraise_exception(): [ 635.570058] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.570058] env[62552]: self.force_reraise() [ 635.570058] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.570058] env[62552]: raise self.value [ 635.570058] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.570058] env[62552]: updated_port = self._update_port( [ 635.570058] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.570818] env[62552]: _ensure_no_port_binding_failure(port) [ 635.570818] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.570818] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.570818] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. [ 635.570818] env[62552]: During handling of the above exception, another exception occurred: [ 635.570818] env[62552]: Traceback (most recent call last): [ 635.570818] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2462, in _do_build_and_run_instance [ 635.570818] env[62552]: self._build_and_run_instance(context, instance, image, [ 635.570818] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2754, in _build_and_run_instance [ 635.570818] env[62552]: raise exception.RescheduledException( [ 635.570818] env[62552]: nova.exception.RescheduledException: Build of instance 92b9d5a7-e5ed-480e-b68e-b589e6d00832 was re-scheduled: Binding failed for port 673852d0-7384-41cc-912f-b25d6edbb3d1, please check neutron logs for more information. [ 635.570818] env[62552]: During handling of the above exception, another exception occurred: [ 635.570818] env[62552]: Traceback (most recent call last): [ 635.570818] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 635.570818] env[62552]: func(*args, **kwargs) [ 635.570818] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.570818] env[62552]: return func(*args, **kwargs) [ 635.570818] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 635.570818] env[62552]: return f(*args, **kwargs) [ 635.570818] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2353, in _locked_do_build_and_run_instance [ 635.570818] env[62552]: result = self._do_build_and_run_instance(*args, **kwargs) [ 635.570818] env[62552]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 635.570818] env[62552]: with excutils.save_and_reraise_exception(): [ 635.570818] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.570818] env[62552]: self.force_reraise() [ 635.570818] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.570818] env[62552]: raise self.value [ 635.570818] env[62552]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 635.572452] env[62552]: return f(self, context, *args, **kw) [ 635.572452] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 168, in decorated_function [ 635.572452] env[62552]: with excutils.save_and_reraise_exception(): [ 635.572452] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.572452] env[62552]: self.force_reraise() [ 635.572452] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.572452] env[62552]: raise self.value [ 635.572452] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 159, in decorated_function [ 635.572452] env[62552]: return function(self, context, *args, **kwargs) [ 635.572452] env[62552]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 635.572452] env[62552]: return function(self, context, *args, **kwargs) [ 635.572452] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 205, in decorated_function [ 635.572452] env[62552]: return function(self, context, *args, **kwargs) [ 635.572452] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2497, in _do_build_and_run_instance [ 635.572452] env[62552]: instance.save() [ 635.572452] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 635.572452] env[62552]: updates, result = self.indirection_api.object_action( [ 635.572452] env[62552]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 635.572452] env[62552]: return cctxt.call(context, 'object_action', objinst=objinst, [ 635.572452] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 635.572452] env[62552]: result = self.transport._send( [ 635.572452] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 635.572452] env[62552]: return self._driver.send(target, ctxt, message, [ 635.572452] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 635.572452] env[62552]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 635.572452] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 635.572452] env[62552]: raise result [ 635.572452] env[62552]: nova.exception_Remote.InstanceNotFound_Remote: Instance 92b9d5a7-e5ed-480e-b68e-b589e6d00832 could not be found. [ 635.572452] env[62552]: Traceback (most recent call last): [ 635.573111] env[62552]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 635.573111] env[62552]: return getattr(target, method)(*args, **kwargs) [ 635.573111] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 635.573111] env[62552]: return fn(self, *args, **kwargs) [ 635.573111] env[62552]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 635.573111] env[62552]: old_ref, inst_ref = db.instance_update_and_get_original( [ 635.573111] env[62552]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 635.573111] env[62552]: return f(*args, **kwargs) [ 635.573111] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 635.573111] env[62552]: with excutils.save_and_reraise_exception() as ectxt: [ 635.573111] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.573111] env[62552]: self.force_reraise() [ 635.573111] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.573111] env[62552]: raise self.value [ 635.573111] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 635.573111] env[62552]: return f(*args, **kwargs) [ 635.573111] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 635.573111] env[62552]: return f(context, *args, **kwargs) [ 635.573111] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 635.573111] env[62552]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 635.573111] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 635.573111] env[62552]: raise exception.InstanceNotFound(instance_id=uuid) [ 635.573111] env[62552]: nova.exception.InstanceNotFound: Instance 92b9d5a7-e5ed-480e-b68e-b589e6d00832 could not be found. [ 635.801837] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.802098] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.802259] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.802439] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.802581] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.802742] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.802938] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.803194] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.803395] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.803588] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.803804] env[62552]: DEBUG nova.virt.hardware [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.804717] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d8202d-d64e-442f-a4ef-661cb37eaa55 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.814424] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d774fa1f-5cd8-41fe-abd1-874a2a007a62 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.822321] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Acquiring lock "ce5d0165-65f1-4505-9c46-1129c56a8913" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.822606] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "ce5d0165-65f1-4505-9c46-1129c56a8913" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.831196] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 635.837194] env[62552]: DEBUG oslo.service.loopingcall [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.837684] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 635.837936] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-169f00a0-cadc-4433-8d31-1171417d7bf0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.850085] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.850574] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 635.853088] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.223s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.868102] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 635.868102] env[62552]: value = "task-1239247" [ 635.868102] env[62552]: _type = "Task" [ 635.868102] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.877889] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239247, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.073307] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 636.365138] env[62552]: DEBUG nova.compute.utils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.369577] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 636.371229] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 636.390701] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239247, 'name': CreateVM_Task, 'duration_secs': 0.257999} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.390967] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 636.391503] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.391699] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.392156] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 636.392910] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63f59c35-68b4-4189-8417-6384bd70121a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.398491] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 636.398491] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528d8c2e-1caf-a111-be16-a47865839d75" [ 636.398491] env[62552]: _type = "Task" [ 636.398491] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.411618] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528d8c2e-1caf-a111-be16-a47865839d75, 'name': SearchDatastore_Task, 'duration_secs': 0.009631} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.412000] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.412290] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 636.412552] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.412824] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.412904] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 636.413220] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-732fdb41-2cbe-448c-be4d-d4d72f7cc08a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.420690] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 636.423018] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 636.423018] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a18c731a-93d5-4e04-99f8-04b14b0e8523 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.426876] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 636.426876] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52630e9a-60dd-6bda-a3ce-fe7d4d116c4e" [ 636.426876] env[62552]: _type = "Task" [ 636.426876] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.432204] env[62552]: DEBUG nova.policy [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '348679098bff40bda47d87a82ad88e42', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '56bd202958c24dd39d20b734d2047ae2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.438928] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52630e9a-60dd-6bda-a3ce-fe7d4d116c4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.602842] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.602842] env[62552]: INFO nova.scheduler.client.report [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Deleted allocations for instance a34ea7fb-bd17-4353-806d-4cd2f5ee169d [ 636.875486] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 636.889730] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Successfully created port: 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.934229] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2398b0ef-a180-4105-9c4b-b553195e587e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.940487] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52630e9a-60dd-6bda-a3ce-fe7d4d116c4e, 'name': SearchDatastore_Task, 'duration_secs': 0.007975} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.942775] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-895b6a97-5b87-4e22-952f-9cc385f0b017 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.950079] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb04e27-5508-4aa4-90af-57f80d183f32 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.953985] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 636.953985] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529835b3-fb0e-cf8f-c253-f9b9b4afa449" [ 636.953985] env[62552]: _type = "Task" [ 636.953985] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.985019] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e6e781-8931-41bf-b37d-9350786f61c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.990864] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529835b3-fb0e-cf8f-c253-f9b9b4afa449, 'name': SearchDatastore_Task, 'duration_secs': 0.008829} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.991515] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.991796] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 636.992069] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-818027fd-dbfa-4977-ba3e-2ede32b8fa66 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.666960] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9753d5b5-10b5-4c10-9786-f77aef71abca tempest-VolumesAssistedSnapshotsTest-2033928810 tempest-VolumesAssistedSnapshotsTest-2033928810-project-member] Lock "a34ea7fb-bd17-4353-806d-4cd2f5ee169d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.113s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.673019] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fd8808-5851-4856-a21b-d5e1b1eacd8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.678932] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 637.678932] env[62552]: value = "task-1239248" [ 637.678932] env[62552]: _type = "Task" [ 637.678932] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.690460] env[62552]: DEBUG nova.compute.provider_tree [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.701152] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239248, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.180656] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 638.183611] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 638.197211] env[62552]: DEBUG nova.scheduler.client.report [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 638.198977] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239248, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.215939] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:31:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1276304853',id=18,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-198845359',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.216210] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.216366] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.216543] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.216689] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.216831] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.217050] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.217212] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.217375] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.217532] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.217698] env[62552]: DEBUG nova.virt.hardware [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.218636] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bbe9bc-6ec7-4029-ac4f-26fafc77fdeb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.227827] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce1ec042-8006-4651-9bcf-ce868fcca627 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.701077] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.847s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.702358] env[62552]: ERROR nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade, please check neutron logs for more information. [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] Traceback (most recent call last): [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self.driver.spawn(context, instance, image_meta, [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] vm_ref = self.build_virtual_machine(instance, [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.702358] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] for vif in network_info: [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] return self._sync_wrapper(fn, *args, **kwargs) [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self.wait() [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self[:] = self._gt.wait() [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] return self._exit_event.wait() [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] result = hub.switch() [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.702617] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] return self.greenlet.switch() [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] result = function(*args, **kwargs) [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] return func(*args, **kwargs) [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] raise e [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] nwinfo = self.network_api.allocate_for_instance( [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] created_port_ids = self._update_ports_for_instance( [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] with excutils.save_and_reraise_exception(): [ 638.702999] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] self.force_reraise() [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] raise self.value [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] updated_port = self._update_port( [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] _ensure_no_port_binding_failure(port) [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] raise exception.PortBindingFailed(port_id=port['id']) [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] nova.exception.PortBindingFailed: Binding failed for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade, please check neutron logs for more information. [ 638.703290] env[62552]: ERROR nova.compute.manager [instance: 58c91f56-a683-4a62-bd45-80c627622621] [ 638.703539] env[62552]: DEBUG nova.compute.utils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Binding failed for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.704848] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Build of instance 58c91f56-a683-4a62-bd45-80c627622621 was re-scheduled: Binding failed for port 70a0ceec-7c96-4714-8d80-d39c5ea54ade, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 638.704848] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 638.704848] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquiring lock "refresh_cache-58c91f56-a683-4a62-bd45-80c627622621" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.705391] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Acquired lock "refresh_cache-58c91f56-a683-4a62-bd45-80c627622621" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.705391] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.716252] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.622s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.719762] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239248, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.603165} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.724623] env[62552]: ERROR nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7, please check neutron logs for more information. [ 638.724623] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.724623] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.724623] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.724623] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.724623] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.724623] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.724623] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.724623] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.724623] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 638.724623] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.724623] env[62552]: ERROR nova.compute.manager raise self.value [ 638.724623] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.724623] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.724623] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.724623] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.725064] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.725064] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.725064] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7, please check neutron logs for more information. [ 638.725064] env[62552]: ERROR nova.compute.manager [ 638.725064] env[62552]: Traceback (most recent call last): [ 638.725064] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.725064] env[62552]: listener.cb(fileno) [ 638.725064] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.725064] env[62552]: result = function(*args, **kwargs) [ 638.725064] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.725064] env[62552]: return func(*args, **kwargs) [ 638.725064] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.725064] env[62552]: raise e [ 638.725064] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.725064] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 638.725064] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.725064] env[62552]: created_port_ids = self._update_ports_for_instance( [ 638.725064] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.725064] env[62552]: with excutils.save_and_reraise_exception(): [ 638.725064] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.725064] env[62552]: self.force_reraise() [ 638.725064] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.725064] env[62552]: raise self.value [ 638.725064] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.725064] env[62552]: updated_port = self._update_port( [ 638.725064] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.725064] env[62552]: _ensure_no_port_binding_failure(port) [ 638.725064] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.725064] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.725823] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7, please check neutron logs for more information. [ 638.725823] env[62552]: Removing descriptor: 16 [ 638.725823] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.725823] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 638.725823] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 638.725823] env[62552]: ERROR nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7, please check neutron logs for more information. [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Traceback (most recent call last): [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] yield resources [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self.driver.spawn(context, instance, image_meta, [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] vm_ref = self.build_virtual_machine(instance, [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.726019] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] for vif in network_info: [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] return self._sync_wrapper(fn, *args, **kwargs) [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self.wait() [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self[:] = self._gt.wait() [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] return self._exit_event.wait() [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] result = hub.switch() [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] return self.greenlet.switch() [ 638.726642] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] result = function(*args, **kwargs) [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] return func(*args, **kwargs) [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] raise e [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] nwinfo = self.network_api.allocate_for_instance( [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] created_port_ids = self._update_ports_for_instance( [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] with excutils.save_and_reraise_exception(): [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.727013] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self.force_reraise() [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] raise self.value [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] updated_port = self._update_port( [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] _ensure_no_port_binding_failure(port) [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] raise exception.PortBindingFailed(port_id=port['id']) [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] nova.exception.PortBindingFailed: Binding failed for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7, please check neutron logs for more information. [ 638.727297] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] [ 638.727297] env[62552]: INFO nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Terminating instance [ 638.728681] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c52aa2d8-0dba-488b-ae7d-8fecd26460b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.736913] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 638.736913] env[62552]: value = "task-1239249" [ 638.736913] env[62552]: _type = "Task" [ 638.736913] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.749253] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239249, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.875101] env[62552]: DEBUG nova.compute.manager [req-30768ba8-3ae0-437a-8f97-c91b5ea45aca req-fbb1646e-fbc5-4d3c-a7a8-acf28a965a2f service nova] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Received event network-changed-9ad52185-b9cb-4624-b9d0-5fa925cd0cf7 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 638.875101] env[62552]: DEBUG nova.compute.manager [req-30768ba8-3ae0-437a-8f97-c91b5ea45aca req-fbb1646e-fbc5-4d3c-a7a8-acf28a965a2f service nova] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Refreshing instance network info cache due to event network-changed-9ad52185-b9cb-4624-b9d0-5fa925cd0cf7. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 638.875101] env[62552]: DEBUG oslo_concurrency.lockutils [req-30768ba8-3ae0-437a-8f97-c91b5ea45aca req-fbb1646e-fbc5-4d3c-a7a8-acf28a965a2f service nova] Acquiring lock "refresh_cache-2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.875101] env[62552]: DEBUG oslo_concurrency.lockutils [req-30768ba8-3ae0-437a-8f97-c91b5ea45aca req-fbb1646e-fbc5-4d3c-a7a8-acf28a965a2f service nova] Acquired lock "refresh_cache-2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.875101] env[62552]: DEBUG nova.network.neutron [req-30768ba8-3ae0-437a-8f97-c91b5ea45aca req-fbb1646e-fbc5-4d3c-a7a8-acf28a965a2f service nova] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Refreshing network info cache for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 639.235263] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Acquiring lock "refresh_cache-2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.244715] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.252276] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239249, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066523} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.252615] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 639.253486] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e0966d-971f-4fd4-a097-f780058a0f54 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.276141] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 639.278911] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1c5428a-6752-4046-8f30-55bca54ae46e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.299103] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 639.299103] env[62552]: value = "task-1239250" [ 639.299103] env[62552]: _type = "Task" [ 639.299103] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.310814] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239250, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.371150] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.401282] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "0418260a-aa27-4955-ab15-b180ec04f0b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.401507] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.402817] env[62552]: DEBUG nova.network.neutron [req-30768ba8-3ae0-437a-8f97-c91b5ea45aca req-fbb1646e-fbc5-4d3c-a7a8-acf28a965a2f service nova] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.474106] env[62552]: DEBUG nova.network.neutron [req-30768ba8-3ae0-437a-8f97-c91b5ea45aca req-fbb1646e-fbc5-4d3c-a7a8-acf28a965a2f service nova] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.663441] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8156d662-2041-4fb7-80b0-54623f452514 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.671252] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1aa4632-3950-478d-8cd0-fc8206a3b701 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.703032] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3c1ea4-a011-465a-88ef-ad61345115ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.710295] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e82570-b8cb-44f8-b29e-76d1f87d7624 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.723578] env[62552]: DEBUG nova.compute.provider_tree [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.808435] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239250, 'name': ReconfigVM_Task, 'duration_secs': 0.291323} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.808697] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440/9952d8f6-ee78-4c2d-b147-8c08c027f440.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 639.809338] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b44071cf-f34d-4623-9a6e-7fe96f726090 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.815209] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 639.815209] env[62552]: value = "task-1239251" [ 639.815209] env[62552]: _type = "Task" [ 639.815209] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.822343] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239251, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.875732] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Releasing lock "refresh_cache-58c91f56-a683-4a62-bd45-80c627622621" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.876043] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 639.876287] env[62552]: DEBUG nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 639.876510] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.896694] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.978974] env[62552]: DEBUG oslo_concurrency.lockutils [req-30768ba8-3ae0-437a-8f97-c91b5ea45aca req-fbb1646e-fbc5-4d3c-a7a8-acf28a965a2f service nova] Releasing lock "refresh_cache-2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.979121] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Acquired lock "refresh_cache-2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.979326] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.227792] env[62552]: DEBUG nova.scheduler.client.report [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 640.327689] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239251, 'name': Rename_Task, 'duration_secs': 0.128075} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.327969] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 640.328451] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac2ba51f-16f9-45dd-ab6b-9102a9c87bcf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.334252] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Waiting for the task: (returnval){ [ 640.334252] env[62552]: value = "task-1239252" [ 640.334252] env[62552]: _type = "Task" [ 640.334252] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.342839] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.401060] env[62552]: DEBUG nova.network.neutron [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.499122] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.574307] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.735610] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.019s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.735610] env[62552]: ERROR nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71, please check neutron logs for more information. [ 640.735610] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Traceback (most recent call last): [ 640.735610] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 640.735610] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self.driver.spawn(context, instance, image_meta, [ 640.735610] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 640.735610] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.735610] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.735610] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] vm_ref = self.build_virtual_machine(instance, [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] for vif in network_info: [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] return self._sync_wrapper(fn, *args, **kwargs) [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self.wait() [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self[:] = self._gt.wait() [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] return self._exit_event.wait() [ 640.736057] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] result = hub.switch() [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] return self.greenlet.switch() [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] result = function(*args, **kwargs) [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] return func(*args, **kwargs) [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] raise e [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] nwinfo = self.network_api.allocate_for_instance( [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.736352] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] created_port_ids = self._update_ports_for_instance( [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] with excutils.save_and_reraise_exception(): [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] self.force_reraise() [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] raise self.value [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] updated_port = self._update_port( [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] _ensure_no_port_binding_failure(port) [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.736641] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] raise exception.PortBindingFailed(port_id=port['id']) [ 640.736905] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] nova.exception.PortBindingFailed: Binding failed for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71, please check neutron logs for more information. [ 640.736905] env[62552]: ERROR nova.compute.manager [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] [ 640.736905] env[62552]: DEBUG nova.compute.utils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Binding failed for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.737738] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.800s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.739083] env[62552]: INFO nova.compute.claims [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.742355] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Build of instance 729eba7d-5f14-4506-8fee-24f6eb685b12 was re-scheduled: Binding failed for port 47679719-5fb5-4c16-a0e2-ca9b2a678e71, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 640.742355] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 640.742355] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Acquiring lock "refresh_cache-729eba7d-5f14-4506-8fee-24f6eb685b12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.742565] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Acquired lock "refresh_cache-729eba7d-5f14-4506-8fee-24f6eb685b12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.742620] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.845545] env[62552]: DEBUG oslo_vmware.api [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Task: {'id': task-1239252, 'name': PowerOnVM_Task, 'duration_secs': 0.423485} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.845852] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 640.846089] env[62552]: DEBUG nova.compute.manager [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 640.846930] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da891b9b-910b-4f4b-94aa-6c2683952371 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.894158] env[62552]: DEBUG nova.compute.manager [req-4849259b-f051-4047-b196-fb36223fb20c req-64679378-0958-4d3a-b5f5-6f278a18e183 service nova] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Received event network-vif-deleted-9ad52185-b9cb-4624-b9d0-5fa925cd0cf7 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 640.903625] env[62552]: INFO nova.compute.manager [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] [instance: 58c91f56-a683-4a62-bd45-80c627622621] Took 1.03 seconds to deallocate network for instance. [ 641.077196] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Releasing lock "refresh_cache-2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.077380] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 641.077576] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.078079] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1784f2f-0ce6-4ab9-acee-a3c4e7758e41 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.087430] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcc28fa-6a42-4e2a-b80e-0146879041e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.110074] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9 could not be found. [ 641.110320] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.110498] env[62552]: INFO nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 641.110743] env[62552]: DEBUG oslo.service.loopingcall [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.110970] env[62552]: DEBUG nova.compute.manager [-] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 641.111072] env[62552]: DEBUG nova.network.neutron [-] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.145246] env[62552]: DEBUG nova.network.neutron [-] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.276407] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.369196] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.456332] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.648842] env[62552]: DEBUG nova.network.neutron [-] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.710394] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "9952d8f6-ee78-4c2d-b147-8c08c027f440" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.710507] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "9952d8f6-ee78-4c2d-b147-8c08c027f440" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.710786] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "9952d8f6-ee78-4c2d-b147-8c08c027f440-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.711009] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "9952d8f6-ee78-4c2d-b147-8c08c027f440-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.711222] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "9952d8f6-ee78-4c2d-b147-8c08c027f440-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.714907] env[62552]: INFO nova.compute.manager [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Terminating instance [ 641.931365] env[62552]: INFO nova.scheduler.client.report [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Deleted allocations for instance 58c91f56-a683-4a62-bd45-80c627622621 [ 641.957687] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Releasing lock "refresh_cache-729eba7d-5f14-4506-8fee-24f6eb685b12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.960983] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 641.960983] env[62552]: DEBUG nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 641.960983] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.978676] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.152084] env[62552]: INFO nova.compute.manager [-] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Took 1.04 seconds to deallocate network for instance. [ 642.154448] env[62552]: DEBUG nova.compute.claims [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 642.154541] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.194508] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09edace2-81c8-4972-b7ae-6da0436aa64e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.203274] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b54076-bf26-4495-b5ca-5cd633d76437 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.238970] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "refresh_cache-9952d8f6-ee78-4c2d-b147-8c08c027f440" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.239254] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquired lock "refresh_cache-9952d8f6-ee78-4c2d-b147-8c08c027f440" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.239513] env[62552]: DEBUG nova.network.neutron [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.241844] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387dfd80-ddac-4337-9fa6-b1ce66f00ca9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.249519] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f17467c-b660-48ff-9cb0-b67d5176a414 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.264163] env[62552]: DEBUG nova.compute.provider_tree [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.449708] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3408ab9b-d5ae-4436-97af-f15e1af798d1 tempest-ServersAdminTestJSON-754110282 tempest-ServersAdminTestJSON-754110282-project-member] Lock "58c91f56-a683-4a62-bd45-80c627622621" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.283s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.483964] env[62552]: DEBUG nova.network.neutron [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.764158] env[62552]: DEBUG nova.network.neutron [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.768076] env[62552]: DEBUG nova.scheduler.client.report [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 642.887985] env[62552]: DEBUG nova.network.neutron [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.953517] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 642.986011] env[62552]: INFO nova.compute.manager [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] [instance: 729eba7d-5f14-4506-8fee-24f6eb685b12] Took 1.03 seconds to deallocate network for instance. [ 643.271070] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.271656] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 643.274373] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.776s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.390367] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Releasing lock "refresh_cache-9952d8f6-ee78-4c2d-b147-8c08c027f440" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.390811] env[62552]: DEBUG nova.compute.manager [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 643.391037] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.391985] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95aedade-6c64-4b89-aebe-e05ba0e9ae51 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.400537] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 643.400811] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51630f61-b5e7-44f0-b289-128135d727a1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.407234] env[62552]: DEBUG oslo_vmware.api [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 643.407234] env[62552]: value = "task-1239253" [ 643.407234] env[62552]: _type = "Task" [ 643.407234] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.416854] env[62552]: DEBUG oslo_vmware.api [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.480411] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.781895] env[62552]: DEBUG nova.compute.utils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.783070] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 643.784023] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 643.883071] env[62552]: DEBUG nova.policy [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '75e562e8ec194b6191f649c726cadd9e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eaca0be8fba14e31a2770fdab508a6aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.920250] env[62552]: DEBUG oslo_vmware.api [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239253, 'name': PowerOffVM_Task, 'duration_secs': 0.18636} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.920536] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 643.920699] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 643.920946] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c47f7174-7484-412b-a272-f41dc7662d1f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.951172] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 643.951420] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 643.951600] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Deleting the datastore file [datastore2] 9952d8f6-ee78-4c2d-b147-8c08c027f440 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 643.951859] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1cdaa70-651e-4951-8a17-fff4181c4059 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.958353] env[62552]: DEBUG oslo_vmware.api [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for the task: (returnval){ [ 643.958353] env[62552]: value = "task-1239255" [ 643.958353] env[62552]: _type = "Task" [ 643.958353] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.967249] env[62552]: DEBUG oslo_vmware.api [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.022828] env[62552]: INFO nova.scheduler.client.report [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Deleted allocations for instance 729eba7d-5f14-4506-8fee-24f6eb685b12 [ 644.279857] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26397f47-f018-42e7-93d7-e546d9fbc4dc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.288590] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 644.298772] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d895a3a3-d4f3-4f76-bae5-6641bee1ad30 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.335643] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b53b421-af3b-4be5-b71a-1f24143d92b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.345054] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272a4d73-0adc-4fbb-9d02-4cff401c3ded {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.358981] env[62552]: DEBUG nova.compute.provider_tree [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.468409] env[62552]: DEBUG oslo_vmware.api [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Task: {'id': task-1239255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09194} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.468709] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 644.468963] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 644.469198] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 644.469410] env[62552]: INFO nova.compute.manager [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Took 1.08 seconds to destroy the instance on the hypervisor. [ 644.469683] env[62552]: DEBUG oslo.service.loopingcall [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.469934] env[62552]: DEBUG nova.compute.manager [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 644.470078] env[62552]: DEBUG nova.network.neutron [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 644.512361] env[62552]: DEBUG nova.network.neutron [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.534488] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d813f61-c74f-470d-a827-46f7779b811a tempest-ImagesNegativeTestJSON-1548353605 tempest-ImagesNegativeTestJSON-1548353605-project-member] Lock "729eba7d-5f14-4506-8fee-24f6eb685b12" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.488s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.669558] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Successfully created port: 1a3e3b55-d40a-4112-a72d-1d48e209ca48 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 644.863984] env[62552]: DEBUG nova.scheduler.client.report [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 645.015745] env[62552]: DEBUG nova.network.neutron [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.035949] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 645.307594] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 645.343305] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 645.343512] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 645.343670] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.343857] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 645.344016] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.345296] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 645.345534] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 645.345833] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 645.345894] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 645.350021] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 645.350021] env[62552]: DEBUG nova.virt.hardware [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.350021] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6d8c90-7c62-4aa7-b1f1-e842959523f2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.356205] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d386fd1c-6b16-4bc7-af03-c7426f3b824f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.376102] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.100s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.376102] env[62552]: ERROR nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c, please check neutron logs for more information. [ 645.376102] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Traceback (most recent call last): [ 645.376102] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 645.376102] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self.driver.spawn(context, instance, image_meta, [ 645.376102] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 645.376102] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.376102] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.376102] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] vm_ref = self.build_virtual_machine(instance, [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] for vif in network_info: [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] return self._sync_wrapper(fn, *args, **kwargs) [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self.wait() [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self[:] = self._gt.wait() [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] return self._exit_event.wait() [ 645.376374] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] result = hub.switch() [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] return self.greenlet.switch() [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] result = function(*args, **kwargs) [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] return func(*args, **kwargs) [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] raise e [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] nwinfo = self.network_api.allocate_for_instance( [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.376712] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] created_port_ids = self._update_ports_for_instance( [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] with excutils.save_and_reraise_exception(): [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] self.force_reraise() [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] raise self.value [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] updated_port = self._update_port( [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] _ensure_no_port_binding_failure(port) [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.376987] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] raise exception.PortBindingFailed(port_id=port['id']) [ 645.377273] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] nova.exception.PortBindingFailed: Binding failed for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c, please check neutron logs for more information. [ 645.377273] env[62552]: ERROR nova.compute.manager [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] [ 645.377273] env[62552]: DEBUG nova.compute.utils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Binding failed for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 645.377365] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.356s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.378764] env[62552]: INFO nova.compute.claims [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.385018] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Build of instance e22c0fef-3d0d-4339-be0b-f129b3628b80 was re-scheduled: Binding failed for port 161e12f6-e3df-4714-baf4-ef7aaa99ed9c, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 645.385018] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 645.385018] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Acquiring lock "refresh_cache-e22c0fef-3d0d-4339-be0b-f129b3628b80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.385018] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Acquired lock "refresh_cache-e22c0fef-3d0d-4339-be0b-f129b3628b80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.385333] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.519617] env[62552]: INFO nova.compute.manager [-] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Took 1.05 seconds to deallocate network for instance. [ 645.555305] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.925851] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.027229] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.177116] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.677801] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Releasing lock "refresh_cache-e22c0fef-3d0d-4339-be0b-f129b3628b80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.678945] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 646.678945] env[62552]: DEBUG nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 646.678945] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.779731] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.868093] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f43f700-c16a-4bd0-963b-c176cde7823e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.876066] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbda7ffc-d779-41d1-9b0d-7ecbff825642 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.911983] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e305450-2770-4fd4-8c5c-a2373892cead {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.920027] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e5af8e-5897-438a-a4e0-642fd9860460 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.935587] env[62552]: DEBUG nova.compute.provider_tree [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.287039] env[62552]: DEBUG nova.network.neutron [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.292181] env[62552]: ERROR nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48, please check neutron logs for more information. [ 647.292181] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 647.292181] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.292181] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 647.292181] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.292181] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 647.292181] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.292181] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 647.292181] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.292181] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 647.292181] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.292181] env[62552]: ERROR nova.compute.manager raise self.value [ 647.292181] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.292181] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 647.292181] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.292181] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 647.292689] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.292689] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 647.292689] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48, please check neutron logs for more information. [ 647.292689] env[62552]: ERROR nova.compute.manager [ 647.292689] env[62552]: Traceback (most recent call last): [ 647.292689] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 647.292689] env[62552]: listener.cb(fileno) [ 647.292689] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.292689] env[62552]: result = function(*args, **kwargs) [ 647.292689] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.292689] env[62552]: return func(*args, **kwargs) [ 647.292689] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 647.292689] env[62552]: raise e [ 647.292689] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.292689] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 647.292689] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.292689] env[62552]: created_port_ids = self._update_ports_for_instance( [ 647.292689] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.292689] env[62552]: with excutils.save_and_reraise_exception(): [ 647.292689] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.292689] env[62552]: self.force_reraise() [ 647.292689] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.292689] env[62552]: raise self.value [ 647.292689] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.292689] env[62552]: updated_port = self._update_port( [ 647.292689] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.292689] env[62552]: _ensure_no_port_binding_failure(port) [ 647.292689] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.292689] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 647.293406] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48, please check neutron logs for more information. [ 647.293406] env[62552]: Removing descriptor: 16 [ 647.293406] env[62552]: ERROR nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48, please check neutron logs for more information. [ 647.293406] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Traceback (most recent call last): [ 647.293406] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 647.293406] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] yield resources [ 647.293406] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 647.293406] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self.driver.spawn(context, instance, image_meta, [ 647.293406] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 647.293406] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.293406] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.293406] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] vm_ref = self.build_virtual_machine(instance, [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] for vif in network_info: [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] return self._sync_wrapper(fn, *args, **kwargs) [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self.wait() [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self[:] = self._gt.wait() [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] return self._exit_event.wait() [ 647.293716] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] result = hub.switch() [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] return self.greenlet.switch() [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] result = function(*args, **kwargs) [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] return func(*args, **kwargs) [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] raise e [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] nwinfo = self.network_api.allocate_for_instance( [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.294041] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] created_port_ids = self._update_ports_for_instance( [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] with excutils.save_and_reraise_exception(): [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self.force_reraise() [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] raise self.value [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] updated_port = self._update_port( [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] _ensure_no_port_binding_failure(port) [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.296869] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] raise exception.PortBindingFailed(port_id=port['id']) [ 647.297227] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] nova.exception.PortBindingFailed: Binding failed for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48, please check neutron logs for more information. [ 647.297227] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] [ 647.297227] env[62552]: INFO nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Terminating instance [ 647.311521] env[62552]: DEBUG nova.compute.manager [req-2f661f10-0f0d-452b-b9dc-3625b601c68f req-c19d8fad-7520-4594-bcca-9de0298b00b5 service nova] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Received event network-changed-1a3e3b55-d40a-4112-a72d-1d48e209ca48 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 647.311761] env[62552]: DEBUG nova.compute.manager [req-2f661f10-0f0d-452b-b9dc-3625b601c68f req-c19d8fad-7520-4594-bcca-9de0298b00b5 service nova] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Refreshing instance network info cache due to event network-changed-1a3e3b55-d40a-4112-a72d-1d48e209ca48. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 647.311930] env[62552]: DEBUG oslo_concurrency.lockutils [req-2f661f10-0f0d-452b-b9dc-3625b601c68f req-c19d8fad-7520-4594-bcca-9de0298b00b5 service nova] Acquiring lock "refresh_cache-b6bff07b-f316-4f26-9623-d50328df1ac0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.312088] env[62552]: DEBUG oslo_concurrency.lockutils [req-2f661f10-0f0d-452b-b9dc-3625b601c68f req-c19d8fad-7520-4594-bcca-9de0298b00b5 service nova] Acquired lock "refresh_cache-b6bff07b-f316-4f26-9623-d50328df1ac0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.312253] env[62552]: DEBUG nova.network.neutron [req-2f661f10-0f0d-452b-b9dc-3625b601c68f req-c19d8fad-7520-4594-bcca-9de0298b00b5 service nova] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Refreshing network info cache for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 647.440019] env[62552]: DEBUG nova.scheduler.client.report [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 647.786240] env[62552]: INFO nova.compute.manager [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] [instance: e22c0fef-3d0d-4339-be0b-f129b3628b80] Took 1.11 seconds to deallocate network for instance. [ 647.797828] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Acquiring lock "refresh_cache-b6bff07b-f316-4f26-9623-d50328df1ac0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.848287] env[62552]: DEBUG nova.network.neutron [req-2f661f10-0f0d-452b-b9dc-3625b601c68f req-c19d8fad-7520-4594-bcca-9de0298b00b5 service nova] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.943986] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.945673] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 647.950821] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.992s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.951422] env[62552]: DEBUG nova.objects.instance [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62552) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 647.972520] env[62552]: DEBUG nova.network.neutron [req-2f661f10-0f0d-452b-b9dc-3625b601c68f req-c19d8fad-7520-4594-bcca-9de0298b00b5 service nova] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.456993] env[62552]: DEBUG nova.compute.utils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.461740] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 648.461925] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 648.478613] env[62552]: DEBUG oslo_concurrency.lockutils [req-2f661f10-0f0d-452b-b9dc-3625b601c68f req-c19d8fad-7520-4594-bcca-9de0298b00b5 service nova] Releasing lock "refresh_cache-b6bff07b-f316-4f26-9623-d50328df1ac0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.478613] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Acquired lock "refresh_cache-b6bff07b-f316-4f26-9623-d50328df1ac0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.478613] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.538219] env[62552]: DEBUG nova.policy [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '384d6877b7384d21b12a155d13056a7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f13a11803d1b44178f053cc0a4affedd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.833833] env[62552]: INFO nova.scheduler.client.report [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Deleted allocations for instance e22c0fef-3d0d-4339-be0b-f129b3628b80 [ 648.965017] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 648.966429] env[62552]: DEBUG oslo_concurrency.lockutils [None req-00d31606-1846-47b8-8dc7-2582cd53001b tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.971835] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.988s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.973857] env[62552]: INFO nova.compute.claims [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.979961] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Successfully created port: 4cbbecb5-081f-453a-9ce0-0462c6f1788e {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.010092] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.189133] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.344294] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b23b9e0c-0481-443e-ab1b-1d7ba434b710 tempest-AttachInterfacesV270Test-1993126778 tempest-AttachInterfacesV270Test-1993126778-project-member] Lock "e22c0fef-3d0d-4339-be0b-f129b3628b80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.579s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.372710] env[62552]: DEBUG nova.compute.manager [req-2ddc852a-34d2-4574-a303-6cefaeafd384 req-377f39f7-5373-4928-8086-14955ca68494 service nova] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Received event network-vif-deleted-1a3e3b55-d40a-4112-a72d-1d48e209ca48 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 649.694342] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Releasing lock "refresh_cache-b6bff07b-f316-4f26-9623-d50328df1ac0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.694342] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 649.694342] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 649.694342] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af02551b-c0b5-44f4-bf95-7b7c727f10e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.706474] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3462ca0d-0dfd-4210-9012-830196dc1c09 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.735475] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6bff07b-f316-4f26-9623-d50328df1ac0 could not be found. [ 649.735862] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.735973] env[62552]: INFO nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 649.736252] env[62552]: DEBUG oslo.service.loopingcall [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.736493] env[62552]: DEBUG nova.compute.manager [-] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 649.736653] env[62552]: DEBUG nova.network.neutron [-] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.771894] env[62552]: DEBUG nova.network.neutron [-] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.847498] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 649.971803] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 650.006562] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.006738] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.006900] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.010732] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.010732] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.010831] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.011063] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.011232] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.011434] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.011554] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.011731] env[62552]: DEBUG nova.virt.hardware [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.012617] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4fe1f9-31fc-4874-b38d-c14918b9b578 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.024717] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4174f465-f314-41bc-847b-033ea5120c80 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.274242] env[62552]: DEBUG nova.network.neutron [-] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.338945] env[62552]: ERROR nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e, please check neutron logs for more information. [ 650.338945] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 650.338945] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.338945] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 650.338945] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.338945] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 650.338945] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.338945] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 650.338945] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.338945] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 650.338945] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.338945] env[62552]: ERROR nova.compute.manager raise self.value [ 650.338945] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.338945] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 650.338945] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.338945] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 650.339413] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.339413] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 650.339413] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e, please check neutron logs for more information. [ 650.339413] env[62552]: ERROR nova.compute.manager [ 650.339413] env[62552]: Traceback (most recent call last): [ 650.339413] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 650.339413] env[62552]: listener.cb(fileno) [ 650.339413] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.339413] env[62552]: result = function(*args, **kwargs) [ 650.339413] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.339413] env[62552]: return func(*args, **kwargs) [ 650.339413] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 650.339413] env[62552]: raise e [ 650.339413] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.339413] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 650.339413] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.339413] env[62552]: created_port_ids = self._update_ports_for_instance( [ 650.339413] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.339413] env[62552]: with excutils.save_and_reraise_exception(): [ 650.339413] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.339413] env[62552]: self.force_reraise() [ 650.339413] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.339413] env[62552]: raise self.value [ 650.339413] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.339413] env[62552]: updated_port = self._update_port( [ 650.339413] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.339413] env[62552]: _ensure_no_port_binding_failure(port) [ 650.339413] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.339413] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 650.340286] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e, please check neutron logs for more information. [ 650.340286] env[62552]: Removing descriptor: 16 [ 650.340286] env[62552]: ERROR nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e, please check neutron logs for more information. [ 650.340286] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Traceback (most recent call last): [ 650.340286] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 650.340286] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] yield resources [ 650.340286] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 650.340286] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self.driver.spawn(context, instance, image_meta, [ 650.340286] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 650.340286] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.340286] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.340286] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] vm_ref = self.build_virtual_machine(instance, [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] for vif in network_info: [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] return self._sync_wrapper(fn, *args, **kwargs) [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self.wait() [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self[:] = self._gt.wait() [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] return self._exit_event.wait() [ 650.340562] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] result = hub.switch() [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] return self.greenlet.switch() [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] result = function(*args, **kwargs) [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] return func(*args, **kwargs) [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] raise e [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] nwinfo = self.network_api.allocate_for_instance( [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.340858] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] created_port_ids = self._update_ports_for_instance( [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] with excutils.save_and_reraise_exception(): [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self.force_reraise() [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] raise self.value [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] updated_port = self._update_port( [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] _ensure_no_port_binding_failure(port) [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.341164] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] raise exception.PortBindingFailed(port_id=port['id']) [ 650.341523] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] nova.exception.PortBindingFailed: Binding failed for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e, please check neutron logs for more information. [ 650.341523] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] [ 650.341523] env[62552]: INFO nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Terminating instance [ 650.374795] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.485487] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1611a67-d861-4fea-bfda-ebe9b626eb9f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.494074] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc1aa1e-424f-4cd6-8b97-94243e176026 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.529651] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c79f7ef-18ba-4e04-bb87-c73fd6961830 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.537092] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-453ba672-db3e-4092-bdb4-6aeedd47e20a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.550741] env[62552]: DEBUG nova.compute.provider_tree [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.778133] env[62552]: INFO nova.compute.manager [-] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Took 1.04 seconds to deallocate network for instance. [ 650.782802] env[62552]: DEBUG nova.compute.claims [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 650.782989] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.845282] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "refresh_cache-7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.846504] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "refresh_cache-7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.846504] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.053974] env[62552]: DEBUG nova.scheduler.client.report [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 651.391205] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.562233] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.592s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.562233] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 651.565658] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.054s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.568041] env[62552]: INFO nova.compute.claims [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 651.608366] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.753376] env[62552]: DEBUG nova.compute.manager [req-feaec105-5849-403c-ac1d-46aa65d789d5 req-26d5365a-b6bf-453c-9441-7a8071deeebd service nova] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Received event network-changed-4cbbecb5-081f-453a-9ce0-0462c6f1788e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 651.753376] env[62552]: DEBUG nova.compute.manager [req-feaec105-5849-403c-ac1d-46aa65d789d5 req-26d5365a-b6bf-453c-9441-7a8071deeebd service nova] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Refreshing instance network info cache due to event network-changed-4cbbecb5-081f-453a-9ce0-0462c6f1788e. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 651.753376] env[62552]: DEBUG oslo_concurrency.lockutils [req-feaec105-5849-403c-ac1d-46aa65d789d5 req-26d5365a-b6bf-453c-9441-7a8071deeebd service nova] Acquiring lock "refresh_cache-7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.082395] env[62552]: DEBUG nova.compute.utils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.089053] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 652.089219] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 652.114346] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "refresh_cache-7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.116605] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 652.116605] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.116605] env[62552]: DEBUG oslo_concurrency.lockutils [req-feaec105-5849-403c-ac1d-46aa65d789d5 req-26d5365a-b6bf-453c-9441-7a8071deeebd service nova] Acquired lock "refresh_cache-7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.116605] env[62552]: DEBUG nova.network.neutron [req-feaec105-5849-403c-ac1d-46aa65d789d5 req-26d5365a-b6bf-453c-9441-7a8071deeebd service nova] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Refreshing network info cache for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 652.116605] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-084bca24-6fb0-46b6-b26b-5e4609389e38 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.128047] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1b5878-c5bd-440e-914d-ea9d27406904 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.152339] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7 could not be found. [ 652.152566] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.152746] env[62552]: INFO nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 652.153011] env[62552]: DEBUG oslo.service.loopingcall [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.153471] env[62552]: DEBUG nova.compute.manager [-] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 652.153569] env[62552]: DEBUG nova.network.neutron [-] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.195387] env[62552]: DEBUG nova.network.neutron [-] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.235731] env[62552]: DEBUG nova.policy [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3eaa776d78b04863937ebf10d840b944', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f6b282bb22f464ea5efc77fc17c3004', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.333380] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "00b06df7-e7ae-48fd-8887-65647931d5cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.333380] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.590374] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 652.652164] env[62552]: DEBUG nova.network.neutron [req-feaec105-5849-403c-ac1d-46aa65d789d5 req-26d5365a-b6bf-453c-9441-7a8071deeebd service nova] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.698340] env[62552]: DEBUG nova.network.neutron [-] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.881405] env[62552]: DEBUG nova.network.neutron [req-feaec105-5849-403c-ac1d-46aa65d789d5 req-26d5365a-b6bf-453c-9441-7a8071deeebd service nova] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.937111] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Successfully created port: 222de83b-1cfc-4f45-b375-c8ca83915c70 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.096551] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc9b092c-0860-4aa4-a10f-9e9fae6917e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.113771] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbc5d81-8e38-48de-ab3c-9713ae148514 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.149263] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aafe715-31c1-4bef-8778-71f6e7a39efa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.157971] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3cab058-7ceb-40ed-96c4-b20737cb66fd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.171898] env[62552]: DEBUG nova.compute.provider_tree [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.201262] env[62552]: INFO nova.compute.manager [-] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Took 1.05 seconds to deallocate network for instance. [ 653.204085] env[62552]: DEBUG nova.compute.claims [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.204483] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.384869] env[62552]: DEBUG oslo_concurrency.lockutils [req-feaec105-5849-403c-ac1d-46aa65d789d5 req-26d5365a-b6bf-453c-9441-7a8071deeebd service nova] Releasing lock "refresh_cache-7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.385159] env[62552]: DEBUG nova.compute.manager [req-feaec105-5849-403c-ac1d-46aa65d789d5 req-26d5365a-b6bf-453c-9441-7a8071deeebd service nova] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Received event network-vif-deleted-4cbbecb5-081f-453a-9ce0-0462c6f1788e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 653.597169] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Successfully created port: 8ea1f1d3-176f-4768-939a-0274baeb93b8 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.608928] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 653.633919] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 653.634189] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 653.634352] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.634528] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 653.634824] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.634824] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 653.635012] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 653.635299] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 653.635475] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 653.635637] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 653.635804] env[62552]: DEBUG nova.virt.hardware [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.636653] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46344450-fd88-4ac8-9324-5d153e8af79b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.644952] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38bc0ff-9346-4650-877e-b950e2bbd4bd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.674582] env[62552]: DEBUG nova.scheduler.client.report [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 653.753650] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 653.753887] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 654.082010] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Successfully created port: 595538e7-0a33-4428-87d4-298acfd753f8 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.179969] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.614s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.180420] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 654.183812] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.583s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.185722] env[62552]: INFO nova.compute.claims [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 654.261679] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 654.263091] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 654.263091] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Rebuilding the list of instances to heal {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 654.686738] env[62552]: DEBUG nova.compute.utils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 654.687839] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 654.687919] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 654.771310] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 654.771615] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 654.771768] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 654.771899] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 654.772075] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 654.772237] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 654.778442] env[62552]: DEBUG nova.policy [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7f8aa8200874dddb71d8b21bd12ca04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8b8d96b464a439e9c7ef6f3e419a9bc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 654.812603] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "refresh_cache-9952d8f6-ee78-4c2d-b147-8c08c027f440" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.812778] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquired lock "refresh_cache-9952d8f6-ee78-4c2d-b147-8c08c027f440" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.812968] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Forcefully refreshing network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 654.813140] env[62552]: DEBUG nova.objects.instance [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lazy-loading 'info_cache' on Instance uuid 9952d8f6-ee78-4c2d-b147-8c08c027f440 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 655.194941] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 655.448498] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Successfully created port: f9f76deb-cd3f-461a-a5ed-313f60107301 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.554275] env[62552]: DEBUG nova.compute.manager [req-e6b2b9b1-68bd-43ba-a59c-48227deaf67d req-3309e855-d60d-47af-8a4f-a6bbffc93dbc service nova] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Received event network-changed-222de83b-1cfc-4f45-b375-c8ca83915c70 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 655.554402] env[62552]: DEBUG nova.compute.manager [req-e6b2b9b1-68bd-43ba-a59c-48227deaf67d req-3309e855-d60d-47af-8a4f-a6bbffc93dbc service nova] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Refreshing instance network info cache due to event network-changed-222de83b-1cfc-4f45-b375-c8ca83915c70. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 655.557069] env[62552]: DEBUG oslo_concurrency.lockutils [req-e6b2b9b1-68bd-43ba-a59c-48227deaf67d req-3309e855-d60d-47af-8a4f-a6bbffc93dbc service nova] Acquiring lock "refresh_cache-bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.557069] env[62552]: DEBUG oslo_concurrency.lockutils [req-e6b2b9b1-68bd-43ba-a59c-48227deaf67d req-3309e855-d60d-47af-8a4f-a6bbffc93dbc service nova] Acquired lock "refresh_cache-bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.557069] env[62552]: DEBUG nova.network.neutron [req-e6b2b9b1-68bd-43ba-a59c-48227deaf67d req-3309e855-d60d-47af-8a4f-a6bbffc93dbc service nova] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Refreshing network info cache for port 222de83b-1cfc-4f45-b375-c8ca83915c70 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 655.679697] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a9afec-0c05-4e13-91cb-1e943c5c13a9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.688402] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce9e7069-024a-4af9-87bd-89757663c1a0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.725784] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72688dad-95bc-4e0f-979c-02d2e17c7907 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.733315] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29425c4-3024-471d-a947-8794dfb006eb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.750598] env[62552]: DEBUG nova.compute.provider_tree [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.876090] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.893015] env[62552]: ERROR nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 222de83b-1cfc-4f45-b375-c8ca83915c70, please check neutron logs for more information. [ 655.893015] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 655.893015] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 655.893015] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 655.893015] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.893015] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 655.893015] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.893015] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 655.893015] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.893015] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 655.893015] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.893015] env[62552]: ERROR nova.compute.manager raise self.value [ 655.893015] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.893015] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 655.893015] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.893015] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 655.893570] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.893570] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 655.893570] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 222de83b-1cfc-4f45-b375-c8ca83915c70, please check neutron logs for more information. [ 655.893570] env[62552]: ERROR nova.compute.manager [ 655.893570] env[62552]: Traceback (most recent call last): [ 655.893570] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 655.893570] env[62552]: listener.cb(fileno) [ 655.893570] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.893570] env[62552]: result = function(*args, **kwargs) [ 655.893570] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.893570] env[62552]: return func(*args, **kwargs) [ 655.893570] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 655.893570] env[62552]: raise e [ 655.893570] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 655.893570] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 655.893570] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.893570] env[62552]: created_port_ids = self._update_ports_for_instance( [ 655.893570] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.893570] env[62552]: with excutils.save_and_reraise_exception(): [ 655.893570] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.893570] env[62552]: self.force_reraise() [ 655.893570] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.893570] env[62552]: raise self.value [ 655.893570] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.893570] env[62552]: updated_port = self._update_port( [ 655.893570] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.893570] env[62552]: _ensure_no_port_binding_failure(port) [ 655.893570] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.893570] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 655.894291] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 222de83b-1cfc-4f45-b375-c8ca83915c70, please check neutron logs for more information. [ 655.894291] env[62552]: Removing descriptor: 16 [ 655.894291] env[62552]: ERROR nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 222de83b-1cfc-4f45-b375-c8ca83915c70, please check neutron logs for more information. [ 655.894291] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Traceback (most recent call last): [ 655.894291] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 655.894291] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] yield resources [ 655.894291] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 655.894291] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self.driver.spawn(context, instance, image_meta, [ 655.894291] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 655.894291] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.894291] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.894291] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] vm_ref = self.build_virtual_machine(instance, [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] for vif in network_info: [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] return self._sync_wrapper(fn, *args, **kwargs) [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self.wait() [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self[:] = self._gt.wait() [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] return self._exit_event.wait() [ 655.894675] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] result = hub.switch() [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] return self.greenlet.switch() [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] result = function(*args, **kwargs) [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] return func(*args, **kwargs) [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] raise e [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] nwinfo = self.network_api.allocate_for_instance( [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.895072] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] created_port_ids = self._update_ports_for_instance( [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] with excutils.save_and_reraise_exception(): [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self.force_reraise() [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] raise self.value [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] updated_port = self._update_port( [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] _ensure_no_port_binding_failure(port) [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.895401] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] raise exception.PortBindingFailed(port_id=port['id']) [ 655.895688] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] nova.exception.PortBindingFailed: Binding failed for port 222de83b-1cfc-4f45-b375-c8ca83915c70, please check neutron logs for more information. [ 655.895688] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] [ 655.895688] env[62552]: INFO nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Terminating instance [ 656.134439] env[62552]: DEBUG nova.network.neutron [req-e6b2b9b1-68bd-43ba-a59c-48227deaf67d req-3309e855-d60d-47af-8a4f-a6bbffc93dbc service nova] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.225938] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 656.254039] env[62552]: DEBUG nova.scheduler.client.report [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 656.274110] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.274358] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.274737] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.274938] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.275097] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.275247] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.275517] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.275607] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.275765] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.275924] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.276116] env[62552]: DEBUG nova.virt.hardware [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.277090] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128da898-66e8-4637-99b5-cc3525369451 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.286594] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8aaa4c-2b0e-4dc3-a69c-b0b5e47e45e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.291590] env[62552]: DEBUG nova.network.neutron [req-e6b2b9b1-68bd-43ba-a59c-48227deaf67d req-3309e855-d60d-47af-8a4f-a6bbffc93dbc service nova] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.399009] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "refresh_cache-bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.488787] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.761769] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.762383] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 656.767429] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.042s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.767683] env[62552]: INFO nova.compute.claims [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.795042] env[62552]: DEBUG oslo_concurrency.lockutils [req-e6b2b9b1-68bd-43ba-a59c-48227deaf67d req-3309e855-d60d-47af-8a4f-a6bbffc93dbc service nova] Releasing lock "refresh_cache-bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.795570] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquired lock "refresh_cache-bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.795775] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.991482] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Releasing lock "refresh_cache-9952d8f6-ee78-4c2d-b147-8c08c027f440" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.994265] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Updated the network info_cache for instance {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 656.994265] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 656.994265] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 656.994265] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 656.994265] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 656.994265] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 656.994637] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 656.994637] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 656.994637] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 657.274563] env[62552]: DEBUG nova.compute.utils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.281043] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 657.281043] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 657.371561] env[62552]: ERROR nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f9f76deb-cd3f-461a-a5ed-313f60107301, please check neutron logs for more information. [ 657.371561] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.371561] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 657.371561] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.371561] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.371561] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.371561] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.371561] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.371561] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.371561] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 657.371561] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.371561] env[62552]: ERROR nova.compute.manager raise self.value [ 657.371561] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.371561] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.371561] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.371561] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.372110] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.372110] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.372110] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f9f76deb-cd3f-461a-a5ed-313f60107301, please check neutron logs for more information. [ 657.372110] env[62552]: ERROR nova.compute.manager [ 657.372110] env[62552]: Traceback (most recent call last): [ 657.372110] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.372110] env[62552]: listener.cb(fileno) [ 657.372110] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.372110] env[62552]: result = function(*args, **kwargs) [ 657.372110] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.372110] env[62552]: return func(*args, **kwargs) [ 657.372110] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 657.372110] env[62552]: raise e [ 657.372110] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 657.372110] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 657.372110] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.372110] env[62552]: created_port_ids = self._update_ports_for_instance( [ 657.372110] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.372110] env[62552]: with excutils.save_and_reraise_exception(): [ 657.372110] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.372110] env[62552]: self.force_reraise() [ 657.372110] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.372110] env[62552]: raise self.value [ 657.372110] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.372110] env[62552]: updated_port = self._update_port( [ 657.372110] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.372110] env[62552]: _ensure_no_port_binding_failure(port) [ 657.372110] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.372110] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.372795] env[62552]: nova.exception.PortBindingFailed: Binding failed for port f9f76deb-cd3f-461a-a5ed-313f60107301, please check neutron logs for more information. [ 657.372795] env[62552]: Removing descriptor: 19 [ 657.372795] env[62552]: ERROR nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f9f76deb-cd3f-461a-a5ed-313f60107301, please check neutron logs for more information. [ 657.372795] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Traceback (most recent call last): [ 657.372795] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 657.372795] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] yield resources [ 657.372795] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 657.372795] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self.driver.spawn(context, instance, image_meta, [ 657.372795] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 657.372795] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.372795] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.372795] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] vm_ref = self.build_virtual_machine(instance, [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] for vif in network_info: [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] return self._sync_wrapper(fn, *args, **kwargs) [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self.wait() [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self[:] = self._gt.wait() [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] return self._exit_event.wait() [ 657.373133] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] result = hub.switch() [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] return self.greenlet.switch() [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] result = function(*args, **kwargs) [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] return func(*args, **kwargs) [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] raise e [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] nwinfo = self.network_api.allocate_for_instance( [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.373474] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] created_port_ids = self._update_ports_for_instance( [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] with excutils.save_and_reraise_exception(): [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self.force_reraise() [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] raise self.value [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] updated_port = self._update_port( [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] _ensure_no_port_binding_failure(port) [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.373808] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] raise exception.PortBindingFailed(port_id=port['id']) [ 657.374135] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] nova.exception.PortBindingFailed: Binding failed for port f9f76deb-cd3f-461a-a5ed-313f60107301, please check neutron logs for more information. [ 657.374135] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] [ 657.374135] env[62552]: INFO nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Terminating instance [ 657.383956] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.424255] env[62552]: DEBUG nova.policy [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d49cc8c015e4c508851cedc2f4ee6e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fb9142d800747a29c6b0ff823a52b3f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 657.496114] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.567873] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.769629] env[62552]: DEBUG nova.compute.manager [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Received event network-vif-deleted-222de83b-1cfc-4f45-b375-c8ca83915c70 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 657.770740] env[62552]: DEBUG nova.compute.manager [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Received event network-changed-f9f76deb-cd3f-461a-a5ed-313f60107301 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 657.770740] env[62552]: DEBUG nova.compute.manager [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Refreshing instance network info cache due to event network-changed-f9f76deb-cd3f-461a-a5ed-313f60107301. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 657.771193] env[62552]: DEBUG oslo_concurrency.lockutils [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] Acquiring lock "refresh_cache-74b58f3f-a575-4790-8a90-cad46b1499e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.772462] env[62552]: DEBUG oslo_concurrency.lockutils [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] Acquired lock "refresh_cache-74b58f3f-a575-4790-8a90-cad46b1499e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.772462] env[62552]: DEBUG nova.network.neutron [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Refreshing network info cache for port f9f76deb-cd3f-461a-a5ed-313f60107301 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 657.784020] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 657.877023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-74b58f3f-a575-4790-8a90-cad46b1499e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.990617] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Successfully created port: 98c5e091-4762-426f-84b0-68293670efc6 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 658.070066] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Releasing lock "refresh_cache-bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.070526] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 658.070725] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 658.071030] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-841b6528-eef7-4f1d-8663-9534f4978227 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.084532] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1957f1a2-f616-48ce-96e3-9c06977bdf6f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.111977] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6 could not be found. [ 658.111977] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 658.111977] env[62552]: INFO nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 658.111977] env[62552]: DEBUG oslo.service.loopingcall [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.112369] env[62552]: DEBUG nova.compute.manager [-] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 658.112471] env[62552]: DEBUG nova.network.neutron [-] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.168099] env[62552]: DEBUG nova.network.neutron [-] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.220257] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36df43be-74a0-48d9-b234-84ee36fe8b49 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.226656] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62739989-d285-4ed2-8700-47d7d9474707 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.261791] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3412ab-82cb-4425-a4b2-17116c93c6a4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.271275] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d3425a-7512-4918-8d1f-fa7b3397e7e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.294175] env[62552]: DEBUG nova.compute.provider_tree [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.311831] env[62552]: DEBUG nova.network.neutron [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.509860] env[62552]: DEBUG nova.network.neutron [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.800208] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 658.802083] env[62552]: DEBUG nova.scheduler.client.report [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 658.838846] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 658.838846] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 658.839147] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 658.839147] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 658.839233] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 658.839662] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 658.839662] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 658.840340] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 658.840340] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 658.840755] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 658.841029] env[62552]: DEBUG nova.virt.hardware [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.842461] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ee2702-1936-4a67-91d2-daaf4bc0afe0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.851758] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c0f7ba8-fafa-46a5-9720-07cc549d8ba4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.012869] env[62552]: DEBUG oslo_concurrency.lockutils [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] Releasing lock "refresh_cache-74b58f3f-a575-4790-8a90-cad46b1499e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.013229] env[62552]: DEBUG nova.compute.manager [req-6c0ab9b9-9970-46b9-ad0a-d98c6ac02606 req-d3f02f7b-061d-4402-8a2c-00cd5ba606c1 service nova] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Received event network-vif-deleted-f9f76deb-cd3f-461a-a5ed-313f60107301 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 659.013649] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-74b58f3f-a575-4790-8a90-cad46b1499e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.013821] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.309525] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.310071] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 659.312615] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.944s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.312797] env[62552]: DEBUG nova.objects.instance [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62552) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 659.543996] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.588579] env[62552]: DEBUG nova.network.neutron [-] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.740252] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.817524] env[62552]: DEBUG nova.compute.utils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.827159] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 659.827159] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 659.947706] env[62552]: DEBUG nova.policy [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae73aca55f7b491a8cd7bd17750dc623', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf213723411c477ea325512314f7aeb0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 660.035520] env[62552]: DEBUG nova.compute.manager [req-3f9a3897-fd19-41ba-9ce2-5a920356b4fd req-033c697c-c8ce-48eb-9cb1-9c69d6b6cce9 service nova] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Received event network-changed-98c5e091-4762-426f-84b0-68293670efc6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 660.035783] env[62552]: DEBUG nova.compute.manager [req-3f9a3897-fd19-41ba-9ce2-5a920356b4fd req-033c697c-c8ce-48eb-9cb1-9c69d6b6cce9 service nova] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Refreshing instance network info cache due to event network-changed-98c5e091-4762-426f-84b0-68293670efc6. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 660.035979] env[62552]: DEBUG oslo_concurrency.lockutils [req-3f9a3897-fd19-41ba-9ce2-5a920356b4fd req-033c697c-c8ce-48eb-9cb1-9c69d6b6cce9 service nova] Acquiring lock "refresh_cache-384445d0-93b0-427c-92aa-ef557351a7cd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.036098] env[62552]: DEBUG oslo_concurrency.lockutils [req-3f9a3897-fd19-41ba-9ce2-5a920356b4fd req-033c697c-c8ce-48eb-9cb1-9c69d6b6cce9 service nova] Acquired lock "refresh_cache-384445d0-93b0-427c-92aa-ef557351a7cd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.036264] env[62552]: DEBUG nova.network.neutron [req-3f9a3897-fd19-41ba-9ce2-5a920356b4fd req-033c697c-c8ce-48eb-9cb1-9c69d6b6cce9 service nova] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Refreshing network info cache for port 98c5e091-4762-426f-84b0-68293670efc6 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 660.045542] env[62552]: ERROR nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 98c5e091-4762-426f-84b0-68293670efc6, please check neutron logs for more information. [ 660.045542] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.045542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 660.045542] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.045542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.045542] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.045542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.045542] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.045542] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.045542] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 660.045542] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.045542] env[62552]: ERROR nova.compute.manager raise self.value [ 660.045542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.045542] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.045542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.045542] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.046625] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.046625] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.046625] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 98c5e091-4762-426f-84b0-68293670efc6, please check neutron logs for more information. [ 660.046625] env[62552]: ERROR nova.compute.manager [ 660.046625] env[62552]: Traceback (most recent call last): [ 660.046625] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.046625] env[62552]: listener.cb(fileno) [ 660.046625] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.046625] env[62552]: result = function(*args, **kwargs) [ 660.046625] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.046625] env[62552]: return func(*args, **kwargs) [ 660.046625] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 660.046625] env[62552]: raise e [ 660.046625] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 660.046625] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 660.046625] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.046625] env[62552]: created_port_ids = self._update_ports_for_instance( [ 660.046625] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.046625] env[62552]: with excutils.save_and_reraise_exception(): [ 660.046625] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.046625] env[62552]: self.force_reraise() [ 660.046625] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.046625] env[62552]: raise self.value [ 660.046625] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.046625] env[62552]: updated_port = self._update_port( [ 660.046625] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.046625] env[62552]: _ensure_no_port_binding_failure(port) [ 660.046625] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.046625] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.047541] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 98c5e091-4762-426f-84b0-68293670efc6, please check neutron logs for more information. [ 660.047541] env[62552]: Removing descriptor: 16 [ 660.047541] env[62552]: ERROR nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 98c5e091-4762-426f-84b0-68293670efc6, please check neutron logs for more information. [ 660.047541] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Traceback (most recent call last): [ 660.047541] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 660.047541] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] yield resources [ 660.047541] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 660.047541] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self.driver.spawn(context, instance, image_meta, [ 660.047541] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 660.047541] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.047541] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.047541] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] vm_ref = self.build_virtual_machine(instance, [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] for vif in network_info: [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] return self._sync_wrapper(fn, *args, **kwargs) [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self.wait() [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self[:] = self._gt.wait() [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] return self._exit_event.wait() [ 660.047847] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] result = hub.switch() [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] return self.greenlet.switch() [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] result = function(*args, **kwargs) [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] return func(*args, **kwargs) [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] raise e [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] nwinfo = self.network_api.allocate_for_instance( [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.048806] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] created_port_ids = self._update_ports_for_instance( [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] with excutils.save_and_reraise_exception(): [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self.force_reraise() [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] raise self.value [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] updated_port = self._update_port( [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] _ensure_no_port_binding_failure(port) [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.049423] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] raise exception.PortBindingFailed(port_id=port['id']) [ 660.050036] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] nova.exception.PortBindingFailed: Binding failed for port 98c5e091-4762-426f-84b0-68293670efc6, please check neutron logs for more information. [ 660.050036] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] [ 660.050036] env[62552]: INFO nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Terminating instance [ 660.091610] env[62552]: INFO nova.compute.manager [-] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Took 1.98 seconds to deallocate network for instance. [ 660.094989] env[62552]: DEBUG nova.compute.claims [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 660.094989] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.176414] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquiring lock "4b80bc41-1b63-444e-8039-696e3d78d9a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.176850] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "4b80bc41-1b63-444e-8039-696e3d78d9a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.243077] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-74b58f3f-a575-4790-8a90-cad46b1499e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.243465] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 660.243653] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 660.243938] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ef9962b-029d-4bff-b2bb-8f332187e6df {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.253278] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc380ad-ff5e-4273-a9d6-730a853456e4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.275041] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 74b58f3f-a575-4790-8a90-cad46b1499e6 could not be found. [ 660.275324] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 660.275523] env[62552]: INFO nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 660.275772] env[62552]: DEBUG oslo.service.loopingcall [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 660.275984] env[62552]: DEBUG nova.compute.manager [-] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 660.276101] env[62552]: DEBUG nova.network.neutron [-] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.321058] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 660.329269] env[62552]: DEBUG oslo_concurrency.lockutils [None req-47adedca-343f-483a-bf09-7ef20dcef7ef tempest-ServersAdmin275Test-109026811 tempest-ServersAdmin275Test-109026811-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.332742] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.176s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.359081] env[62552]: DEBUG nova.network.neutron [-] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.546354] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Successfully created port: 700526b4-3e5f-4f5f-8f32-0b6cb2778723 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.556301] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Acquiring lock "refresh_cache-384445d0-93b0-427c-92aa-ef557351a7cd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.579455] env[62552]: DEBUG nova.network.neutron [req-3f9a3897-fd19-41ba-9ce2-5a920356b4fd req-033c697c-c8ce-48eb-9cb1-9c69d6b6cce9 service nova] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.702264] env[62552]: DEBUG nova.network.neutron [req-3f9a3897-fd19-41ba-9ce2-5a920356b4fd req-033c697c-c8ce-48eb-9cb1-9c69d6b6cce9 service nova] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.860457] env[62552]: DEBUG nova.network.neutron [-] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.205708] env[62552]: DEBUG oslo_concurrency.lockutils [req-3f9a3897-fd19-41ba-9ce2-5a920356b4fd req-033c697c-c8ce-48eb-9cb1-9c69d6b6cce9 service nova] Releasing lock "refresh_cache-384445d0-93b0-427c-92aa-ef557351a7cd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.213261] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Acquired lock "refresh_cache-384445d0-93b0-427c-92aa-ef557351a7cd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.213505] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 661.331850] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 661.334534] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2951b418-b9ac-408d-aaf5-72622fedb499 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.343311] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3b193b-9b43-41cc-913e-ca883b167b55 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.375946] env[62552]: INFO nova.compute.manager [-] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Took 1.10 seconds to deallocate network for instance. [ 661.380126] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae94aca8-88e9-4f15-8513-130255d9b6c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.383086] env[62552]: DEBUG nova.compute.claims [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 661.383268] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.390585] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:31:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='73f50498-07df-4e81-8f0f-cbf4d0fe4bdb',id=21,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1879263063',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.390857] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.391018] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.391219] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.391357] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.391921] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.392106] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.392503] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.392692] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.392854] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.393034] env[62552]: DEBUG nova.virt.hardware [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.395157] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c27ab70-8819-4b6f-8a43-9888d89fa209 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.400020] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59146107-d311-4690-a775-4de238185a24 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.413290] env[62552]: DEBUG nova.compute.provider_tree [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.418079] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6696c930-6d60-4718-bafc-d12fbe86ba87 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.746852] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.866188] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.917990] env[62552]: DEBUG nova.scheduler.client.report [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 662.000408] env[62552]: ERROR nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723, please check neutron logs for more information. [ 662.000408] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 662.000408] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.000408] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 662.000408] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.000408] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 662.000408] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.000408] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 662.000408] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.000408] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 662.000408] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.000408] env[62552]: ERROR nova.compute.manager raise self.value [ 662.000408] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.000408] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 662.000408] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.000408] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 662.000888] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.000888] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 662.000888] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723, please check neutron logs for more information. [ 662.000888] env[62552]: ERROR nova.compute.manager [ 662.000888] env[62552]: Traceback (most recent call last): [ 662.000888] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 662.000888] env[62552]: listener.cb(fileno) [ 662.000888] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.000888] env[62552]: result = function(*args, **kwargs) [ 662.000888] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.000888] env[62552]: return func(*args, **kwargs) [ 662.000888] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 662.000888] env[62552]: raise e [ 662.000888] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.000888] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 662.000888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.000888] env[62552]: created_port_ids = self._update_ports_for_instance( [ 662.000888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.000888] env[62552]: with excutils.save_and_reraise_exception(): [ 662.000888] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.000888] env[62552]: self.force_reraise() [ 662.000888] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.000888] env[62552]: raise self.value [ 662.000888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.000888] env[62552]: updated_port = self._update_port( [ 662.000888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.000888] env[62552]: _ensure_no_port_binding_failure(port) [ 662.000888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.000888] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 662.001580] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723, please check neutron logs for more information. [ 662.001580] env[62552]: Removing descriptor: 15 [ 662.001580] env[62552]: ERROR nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723, please check neutron logs for more information. [ 662.001580] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Traceback (most recent call last): [ 662.001580] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 662.001580] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] yield resources [ 662.001580] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 662.001580] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self.driver.spawn(context, instance, image_meta, [ 662.001580] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 662.001580] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.001580] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.001580] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] vm_ref = self.build_virtual_machine(instance, [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] for vif in network_info: [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] return self._sync_wrapper(fn, *args, **kwargs) [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self.wait() [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self[:] = self._gt.wait() [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] return self._exit_event.wait() [ 662.002297] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] result = hub.switch() [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] return self.greenlet.switch() [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] result = function(*args, **kwargs) [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] return func(*args, **kwargs) [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] raise e [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] nwinfo = self.network_api.allocate_for_instance( [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.002642] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] created_port_ids = self._update_ports_for_instance( [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] with excutils.save_and_reraise_exception(): [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self.force_reraise() [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] raise self.value [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] updated_port = self._update_port( [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] _ensure_no_port_binding_failure(port) [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.002940] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] raise exception.PortBindingFailed(port_id=port['id']) [ 662.003227] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] nova.exception.PortBindingFailed: Binding failed for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723, please check neutron logs for more information. [ 662.003227] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] [ 662.003227] env[62552]: INFO nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Terminating instance [ 662.082322] env[62552]: DEBUG nova.compute.manager [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Received event network-vif-deleted-98c5e091-4762-426f-84b0-68293670efc6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 662.082322] env[62552]: DEBUG nova.compute.manager [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Received event network-changed-700526b4-3e5f-4f5f-8f32-0b6cb2778723 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 662.082322] env[62552]: DEBUG nova.compute.manager [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Refreshing instance network info cache due to event network-changed-700526b4-3e5f-4f5f-8f32-0b6cb2778723. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 662.082322] env[62552]: DEBUG oslo_concurrency.lockutils [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] Acquiring lock "refresh_cache-fae607f4-e18a-4cd4-8843-246a3341a548" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.083026] env[62552]: DEBUG oslo_concurrency.lockutils [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] Acquired lock "refresh_cache-fae607f4-e18a-4cd4-8843-246a3341a548" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.083988] env[62552]: DEBUG nova.network.neutron [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Refreshing network info cache for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 662.369171] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Releasing lock "refresh_cache-384445d0-93b0-427c-92aa-ef557351a7cd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.369622] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 662.373242] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 662.373242] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8ff5040-b5d9-44a2-ae9c-b81442f37bd7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.383403] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44cfdc2-a3d3-4a3d-840e-28ee0e8a410e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.406551] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 384445d0-93b0-427c-92aa-ef557351a7cd could not be found. [ 662.406756] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 662.406908] env[62552]: INFO nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 662.407333] env[62552]: DEBUG oslo.service.loopingcall [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.407555] env[62552]: DEBUG nova.compute.manager [-] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 662.407652] env[62552]: DEBUG nova.network.neutron [-] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 662.422316] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.092s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.423026] env[62552]: ERROR nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7, please check neutron logs for more information. [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Traceback (most recent call last): [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self.driver.spawn(context, instance, image_meta, [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] vm_ref = self.build_virtual_machine(instance, [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.423026] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] for vif in network_info: [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] return self._sync_wrapper(fn, *args, **kwargs) [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self.wait() [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self[:] = self._gt.wait() [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] return self._exit_event.wait() [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] result = hub.switch() [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.423367] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] return self.greenlet.switch() [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] result = function(*args, **kwargs) [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] return func(*args, **kwargs) [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] raise e [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] nwinfo = self.network_api.allocate_for_instance( [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] created_port_ids = self._update_ports_for_instance( [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] with excutils.save_and_reraise_exception(): [ 662.423671] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] self.force_reraise() [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] raise self.value [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] updated_port = self._update_port( [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] _ensure_no_port_binding_failure(port) [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] raise exception.PortBindingFailed(port_id=port['id']) [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] nova.exception.PortBindingFailed: Binding failed for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7, please check neutron logs for more information. [ 662.423967] env[62552]: ERROR nova.compute.manager [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] [ 662.424247] env[62552]: DEBUG nova.compute.utils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Binding failed for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 662.425392] env[62552]: DEBUG nova.network.neutron [-] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.427253] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.947s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.428593] env[62552]: INFO nova.compute.claims [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.434593] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Build of instance 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9 was re-scheduled: Binding failed for port 9ad52185-b9cb-4624-b9d0-5fa925cd0cf7, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 662.434593] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 662.434795] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Acquiring lock "refresh_cache-2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.434834] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Acquired lock "refresh_cache-2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.434980] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 662.505770] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "refresh_cache-fae607f4-e18a-4cd4-8843-246a3341a548" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.619197] env[62552]: DEBUG nova.network.neutron [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.702169] env[62552]: DEBUG nova.network.neutron [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.928855] env[62552]: DEBUG nova.network.neutron [-] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.963740] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.079714] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.205151] env[62552]: DEBUG oslo_concurrency.lockutils [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] Releasing lock "refresh_cache-fae607f4-e18a-4cd4-8843-246a3341a548" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.205441] env[62552]: DEBUG nova.compute.manager [req-d3718ab6-170b-4176-b003-99b528ca7d3e req-32b95934-aeb6-4f25-828f-0804ce2183d4 service nova] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Received event network-vif-deleted-700526b4-3e5f-4f5f-8f32-0b6cb2778723 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 663.206689] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquired lock "refresh_cache-fae607f4-e18a-4cd4-8843-246a3341a548" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.206815] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.433301] env[62552]: INFO nova.compute.manager [-] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Took 1.03 seconds to deallocate network for instance. [ 663.436437] env[62552]: DEBUG nova.compute.claims [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 663.436672] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.477614] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.477863] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.582723] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Releasing lock "refresh_cache-2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.582961] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 663.583161] env[62552]: DEBUG nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 663.583354] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.599938] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.734814] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.805840] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.855715] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47f28ed-8b27-4e96-85f6-7db9835a0c1a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.863838] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d248b50d-bae3-44fe-8302-bddaebb9d5c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.903230] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5a8a64-6769-4657-a759-1c7d32728e11 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.910687] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda3dd04-8c6e-426d-91cf-26e061d30662 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.924608] env[62552]: DEBUG nova.compute.provider_tree [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.104418] env[62552]: DEBUG nova.network.neutron [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.311021] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Releasing lock "refresh_cache-fae607f4-e18a-4cd4-8843-246a3341a548" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.311146] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 664.311345] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 664.311654] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17c54adc-2099-41c7-978f-3701c5ed0040 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.320843] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a3e9c5-fa2c-4510-9a01-2dd5c408ad88 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.341375] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fae607f4-e18a-4cd4-8843-246a3341a548 could not be found. [ 664.341589] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.341771] env[62552]: INFO nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Took 0.03 seconds to destroy the instance on the hypervisor. [ 664.342022] env[62552]: DEBUG oslo.service.loopingcall [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.342224] env[62552]: DEBUG nova.compute.manager [-] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 664.342539] env[62552]: DEBUG nova.network.neutron [-] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.357686] env[62552]: DEBUG nova.network.neutron [-] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.427803] env[62552]: DEBUG nova.scheduler.client.report [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 664.607603] env[62552]: INFO nova.compute.manager [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] [instance: 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9] Took 1.02 seconds to deallocate network for instance. [ 664.860162] env[62552]: DEBUG nova.network.neutron [-] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.933036] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.933490] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 664.936322] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.381s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.937653] env[62552]: INFO nova.compute.claims [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.362786] env[62552]: INFO nova.compute.manager [-] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Took 1.02 seconds to deallocate network for instance. [ 665.365171] env[62552]: DEBUG nova.compute.claims [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 665.365900] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.446332] env[62552]: DEBUG nova.compute.utils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 665.447932] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 665.448083] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 665.506750] env[62552]: DEBUG nova.policy [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12d1227fd454d3a9036c2e6c110a85f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd93b575f2e04fc9b93319545fa03708', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 665.645122] env[62552]: INFO nova.scheduler.client.report [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Deleted allocations for instance 2e904bbe-1dcf-4790-b180-e8d20f9b5bb9 [ 665.834632] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Successfully created port: 43393080-f145-4d14-8609-4c60eb83b25a {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.953113] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 666.154365] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2603da5e-70ba-47a8-8072-ad9e052e5f7b tempest-ServersWithSpecificFlavorTestJSON-1504041818 tempest-ServersWithSpecificFlavorTestJSON-1504041818-project-member] Lock "2e904bbe-1dcf-4790-b180-e8d20f9b5bb9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.574s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.416079] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3362c2-4850-435a-81b0-047cde19d062 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.424023] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecee8dc-abad-4241-8246-cd41edd3af95 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.457290] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa3f675-de39-4b16-9e2e-4e8606dc2025 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.469412] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727f28d2-b8cd-4085-8e55-a4e8dce547bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.486069] env[62552]: DEBUG nova.compute.provider_tree [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.657951] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 666.966966] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 666.992339] env[62552]: DEBUG nova.scheduler.client.report [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 667.000317] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 667.000572] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 667.000727] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 667.000931] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 667.001533] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 667.004970] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 667.004970] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 667.004970] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 667.004970] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 667.004970] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 667.005350] env[62552]: DEBUG nova.virt.hardware [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 667.005350] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533600dd-177c-4b7c-9501-d75c8a27281d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.016276] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9e8d2f-0f29-4ea8-bb9d-c034a9dbe5c3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.114076] env[62552]: DEBUG nova.compute.manager [req-f8f3f601-fe42-49d1-b00b-17979dccb68e req-5e50d600-6d62-4051-8000-9b2308f0e5fa service nova] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Received event network-changed-43393080-f145-4d14-8609-4c60eb83b25a {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 667.114294] env[62552]: DEBUG nova.compute.manager [req-f8f3f601-fe42-49d1-b00b-17979dccb68e req-5e50d600-6d62-4051-8000-9b2308f0e5fa service nova] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Refreshing instance network info cache due to event network-changed-43393080-f145-4d14-8609-4c60eb83b25a. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 667.114508] env[62552]: DEBUG oslo_concurrency.lockutils [req-f8f3f601-fe42-49d1-b00b-17979dccb68e req-5e50d600-6d62-4051-8000-9b2308f0e5fa service nova] Acquiring lock "refresh_cache-187dd75d-cdb0-4b86-a396-f43778fdcc21" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.114692] env[62552]: DEBUG oslo_concurrency.lockutils [req-f8f3f601-fe42-49d1-b00b-17979dccb68e req-5e50d600-6d62-4051-8000-9b2308f0e5fa service nova] Acquired lock "refresh_cache-187dd75d-cdb0-4b86-a396-f43778fdcc21" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.114809] env[62552]: DEBUG nova.network.neutron [req-f8f3f601-fe42-49d1-b00b-17979dccb68e req-5e50d600-6d62-4051-8000-9b2308f0e5fa service nova] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Refreshing network info cache for port 43393080-f145-4d14-8609-4c60eb83b25a {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 667.185509] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.307360] env[62552]: ERROR nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 43393080-f145-4d14-8609-4c60eb83b25a, please check neutron logs for more information. [ 667.307360] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 667.307360] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.307360] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 667.307360] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.307360] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 667.307360] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.307360] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 667.307360] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.307360] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 667.307360] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.307360] env[62552]: ERROR nova.compute.manager raise self.value [ 667.307360] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.307360] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 667.307360] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.307360] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 667.307800] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.307800] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 667.307800] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 43393080-f145-4d14-8609-4c60eb83b25a, please check neutron logs for more information. [ 667.307800] env[62552]: ERROR nova.compute.manager [ 667.307800] env[62552]: Traceback (most recent call last): [ 667.307800] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 667.307800] env[62552]: listener.cb(fileno) [ 667.307800] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.307800] env[62552]: result = function(*args, **kwargs) [ 667.307800] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.307800] env[62552]: return func(*args, **kwargs) [ 667.307800] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.307800] env[62552]: raise e [ 667.307800] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.307800] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 667.307800] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.307800] env[62552]: created_port_ids = self._update_ports_for_instance( [ 667.307800] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.307800] env[62552]: with excutils.save_and_reraise_exception(): [ 667.307800] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.307800] env[62552]: self.force_reraise() [ 667.307800] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.307800] env[62552]: raise self.value [ 667.307800] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.307800] env[62552]: updated_port = self._update_port( [ 667.307800] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.307800] env[62552]: _ensure_no_port_binding_failure(port) [ 667.307800] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.307800] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 667.308562] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 43393080-f145-4d14-8609-4c60eb83b25a, please check neutron logs for more information. [ 667.308562] env[62552]: Removing descriptor: 15 [ 667.308562] env[62552]: ERROR nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 43393080-f145-4d14-8609-4c60eb83b25a, please check neutron logs for more information. [ 667.308562] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Traceback (most recent call last): [ 667.308562] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 667.308562] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] yield resources [ 667.308562] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 667.308562] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self.driver.spawn(context, instance, image_meta, [ 667.308562] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 667.308562] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.308562] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.308562] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] vm_ref = self.build_virtual_machine(instance, [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] for vif in network_info: [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] return self._sync_wrapper(fn, *args, **kwargs) [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self.wait() [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self[:] = self._gt.wait() [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] return self._exit_event.wait() [ 667.308867] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] result = hub.switch() [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] return self.greenlet.switch() [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] result = function(*args, **kwargs) [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] return func(*args, **kwargs) [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] raise e [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] nwinfo = self.network_api.allocate_for_instance( [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.309228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] created_port_ids = self._update_ports_for_instance( [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] with excutils.save_and_reraise_exception(): [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self.force_reraise() [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] raise self.value [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] updated_port = self._update_port( [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] _ensure_no_port_binding_failure(port) [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.309551] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] raise exception.PortBindingFailed(port_id=port['id']) [ 667.309865] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] nova.exception.PortBindingFailed: Binding failed for port 43393080-f145-4d14-8609-4c60eb83b25a, please check neutron logs for more information. [ 667.309865] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] [ 667.309865] env[62552]: INFO nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Terminating instance [ 667.500025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.500147] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 667.503495] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.477s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.503743] env[62552]: DEBUG nova.objects.instance [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lazy-loading 'resources' on Instance uuid 9952d8f6-ee78-4c2d-b147-8c08c027f440 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 667.636622] env[62552]: DEBUG nova.network.neutron [req-f8f3f601-fe42-49d1-b00b-17979dccb68e req-5e50d600-6d62-4051-8000-9b2308f0e5fa service nova] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.814315] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "refresh_cache-187dd75d-cdb0-4b86-a396-f43778fdcc21" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.820615] env[62552]: DEBUG nova.network.neutron [req-f8f3f601-fe42-49d1-b00b-17979dccb68e req-5e50d600-6d62-4051-8000-9b2308f0e5fa service nova] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.005914] env[62552]: DEBUG nova.compute.utils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.009735] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 668.009735] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 668.141022] env[62552]: DEBUG nova.policy [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18603f735ad04cbfa2cc943ab771c7bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '77031446ea2a490ca8e350bdce5ca6c9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.323159] env[62552]: DEBUG oslo_concurrency.lockutils [req-f8f3f601-fe42-49d1-b00b-17979dccb68e req-5e50d600-6d62-4051-8000-9b2308f0e5fa service nova] Releasing lock "refresh_cache-187dd75d-cdb0-4b86-a396-f43778fdcc21" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.323712] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquired lock "refresh_cache-187dd75d-cdb0-4b86-a396-f43778fdcc21" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.323905] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.460070] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9258bf11-ce19-4f8b-9f8f-d049b3ceb25e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.469685] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6015e3-6d35-47c0-b637-4b8edd8ba09d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.503789] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1c62e6-d892-40dc-a3f9-a4d850875310 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.512015] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cba739-89e5-48dd-aca2-d5f1f7fdcbc8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.516438] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 668.528596] env[62552]: DEBUG nova.compute.provider_tree [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.780495] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Successfully created port: 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.879710] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.031331] env[62552]: DEBUG nova.scheduler.client.report [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 669.246222] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.379066] env[62552]: DEBUG nova.compute.manager [req-5ea4a226-02df-4518-a5b7-4b3b615ed35a req-dbc8b48f-e6df-42fd-b943-44afe05fdcf1 service nova] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Received event network-vif-deleted-43393080-f145-4d14-8609-4c60eb83b25a {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 669.532259] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 669.536412] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.033s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.538720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.164s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.540273] env[62552]: INFO nova.compute.claims [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 669.568343] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.568582] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.568738] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.570132] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.570444] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.570955] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.570955] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.570955] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.571132] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.571289] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.571464] env[62552]: DEBUG nova.virt.hardware [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.573770] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d8641a-e0a5-4e24-941c-9f1c4f0b184a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.579035] env[62552]: INFO nova.scheduler.client.report [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Deleted allocations for instance 9952d8f6-ee78-4c2d-b147-8c08c027f440 [ 669.586437] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e48cc8-ffce-4a8e-8458-3e1d685599b3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.754720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Releasing lock "refresh_cache-187dd75d-cdb0-4b86-a396-f43778fdcc21" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.754720] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 669.754720] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 669.754720] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4682f8ac-9943-4ccb-b8d4-49b3fcb26389 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.764042] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ee500d-186c-425b-b0ef-0a433fb712ac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.789792] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 187dd75d-cdb0-4b86-a396-f43778fdcc21 could not be found. [ 669.790046] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 669.790237] env[62552]: INFO nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Took 0.04 seconds to destroy the instance on the hypervisor. [ 669.790482] env[62552]: DEBUG oslo.service.loopingcall [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.790704] env[62552]: DEBUG nova.compute.manager [-] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 669.790801] env[62552]: DEBUG nova.network.neutron [-] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 669.810149] env[62552]: DEBUG nova.network.neutron [-] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.088475] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f063955-58bf-40c0-9ec7-d0e2e86c2663 tempest-ServersAdmin275Test-114797336 tempest-ServersAdmin275Test-114797336-project-member] Lock "9952d8f6-ee78-4c2d-b147-8c08c027f440" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.378s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.314651] env[62552]: DEBUG nova.network.neutron [-] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.358900] env[62552]: ERROR nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8, please check neutron logs for more information. [ 670.358900] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 670.358900] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 670.358900] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 670.358900] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.358900] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 670.358900] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.358900] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 670.358900] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.358900] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 670.358900] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.358900] env[62552]: ERROR nova.compute.manager raise self.value [ 670.358900] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.358900] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 670.358900] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.358900] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 670.359394] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.359394] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 670.359394] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8, please check neutron logs for more information. [ 670.359394] env[62552]: ERROR nova.compute.manager [ 670.359394] env[62552]: Traceback (most recent call last): [ 670.359394] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 670.359394] env[62552]: listener.cb(fileno) [ 670.359394] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.359394] env[62552]: result = function(*args, **kwargs) [ 670.359394] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.359394] env[62552]: return func(*args, **kwargs) [ 670.359394] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 670.359394] env[62552]: raise e [ 670.359394] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 670.359394] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 670.359394] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.359394] env[62552]: created_port_ids = self._update_ports_for_instance( [ 670.359394] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.359394] env[62552]: with excutils.save_and_reraise_exception(): [ 670.359394] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.359394] env[62552]: self.force_reraise() [ 670.359394] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.359394] env[62552]: raise self.value [ 670.359394] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.359394] env[62552]: updated_port = self._update_port( [ 670.359394] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.359394] env[62552]: _ensure_no_port_binding_failure(port) [ 670.359394] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.359394] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 670.360179] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8, please check neutron logs for more information. [ 670.360179] env[62552]: Removing descriptor: 15 [ 670.360179] env[62552]: ERROR nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8, please check neutron logs for more information. [ 670.360179] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Traceback (most recent call last): [ 670.360179] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 670.360179] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] yield resources [ 670.360179] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 670.360179] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self.driver.spawn(context, instance, image_meta, [ 670.360179] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 670.360179] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.360179] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.360179] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] vm_ref = self.build_virtual_machine(instance, [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] for vif in network_info: [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] return self._sync_wrapper(fn, *args, **kwargs) [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self.wait() [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self[:] = self._gt.wait() [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] return self._exit_event.wait() [ 670.360522] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] result = hub.switch() [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] return self.greenlet.switch() [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] result = function(*args, **kwargs) [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] return func(*args, **kwargs) [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] raise e [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] nwinfo = self.network_api.allocate_for_instance( [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.360949] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] created_port_ids = self._update_ports_for_instance( [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] with excutils.save_and_reraise_exception(): [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self.force_reraise() [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] raise self.value [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] updated_port = self._update_port( [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] _ensure_no_port_binding_failure(port) [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.361316] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] raise exception.PortBindingFailed(port_id=port['id']) [ 670.361648] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] nova.exception.PortBindingFailed: Binding failed for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8, please check neutron logs for more information. [ 670.361648] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] [ 670.361648] env[62552]: INFO nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Terminating instance [ 670.818169] env[62552]: INFO nova.compute.manager [-] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Took 1.03 seconds to deallocate network for instance. [ 670.826377] env[62552]: DEBUG nova.compute.claims [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 670.826578] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.865249] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "refresh_cache-1b326e57-365e-49cc-b6f0-eed559ce32f5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.868071] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquired lock "refresh_cache-1b326e57-365e-49cc-b6f0-eed559ce32f5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.868071] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.017050] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b36b169-773a-425b-9e36-d8d8a3bde601 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.028403] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fa1483-49f4-440b-bda3-96d4d8cdfccb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.064530] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2bd5df-2b0f-4640-b554-4fe0e3528d3a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.075771] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07bb2f27-0a80-4e35-b81b-9081370374f7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.091773] env[62552]: DEBUG nova.compute.provider_tree [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.394936] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.428474] env[62552]: DEBUG nova.compute.manager [req-039d4772-b0b8-4e07-b0bb-987657a5a999 req-0dd01cea-4aa0-4789-8e58-3c4982db37e4 service nova] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Received event network-changed-6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 671.428682] env[62552]: DEBUG nova.compute.manager [req-039d4772-b0b8-4e07-b0bb-987657a5a999 req-0dd01cea-4aa0-4789-8e58-3c4982db37e4 service nova] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Refreshing instance network info cache due to event network-changed-6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 671.428875] env[62552]: DEBUG oslo_concurrency.lockutils [req-039d4772-b0b8-4e07-b0bb-987657a5a999 req-0dd01cea-4aa0-4789-8e58-3c4982db37e4 service nova] Acquiring lock "refresh_cache-1b326e57-365e-49cc-b6f0-eed559ce32f5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.460762] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.594886] env[62552]: DEBUG nova.scheduler.client.report [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 671.962585] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Releasing lock "refresh_cache-1b326e57-365e-49cc-b6f0-eed559ce32f5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.964419] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 671.964656] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 671.964984] env[62552]: DEBUG oslo_concurrency.lockutils [req-039d4772-b0b8-4e07-b0bb-987657a5a999 req-0dd01cea-4aa0-4789-8e58-3c4982db37e4 service nova] Acquired lock "refresh_cache-1b326e57-365e-49cc-b6f0-eed559ce32f5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.965250] env[62552]: DEBUG nova.network.neutron [req-039d4772-b0b8-4e07-b0bb-987657a5a999 req-0dd01cea-4aa0-4789-8e58-3c4982db37e4 service nova] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Refreshing network info cache for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 671.966358] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89f490be-91ed-46e7-9e8d-fedb49b9bf07 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.976884] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736d91a9-2f3d-4532-ba00-c04db5ba19a4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.999860] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1b326e57-365e-49cc-b6f0-eed559ce32f5 could not be found. [ 672.000115] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 672.000431] env[62552]: INFO nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 672.000556] env[62552]: DEBUG oslo.service.loopingcall [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.000748] env[62552]: DEBUG nova.compute.manager [-] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 672.000860] env[62552]: DEBUG nova.network.neutron [-] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.020502] env[62552]: DEBUG nova.network.neutron [-] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.102229] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.103274] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 672.105694] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.323s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.491819] env[62552]: DEBUG nova.network.neutron [req-039d4772-b0b8-4e07-b0bb-987657a5a999 req-0dd01cea-4aa0-4789-8e58-3c4982db37e4 service nova] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.522276] env[62552]: DEBUG nova.network.neutron [-] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.572968] env[62552]: DEBUG nova.network.neutron [req-039d4772-b0b8-4e07-b0bb-987657a5a999 req-0dd01cea-4aa0-4789-8e58-3c4982db37e4 service nova] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.610774] env[62552]: DEBUG nova.compute.utils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 672.617021] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 672.617021] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 672.667748] env[62552]: DEBUG nova.policy [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '81ff146ac33b4a5ab1b91df0614c1cfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '173a8b2a661143bb947461a872b418e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.024722] env[62552]: INFO nova.compute.manager [-] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Took 1.02 seconds to deallocate network for instance. [ 673.029691] env[62552]: DEBUG nova.compute.claims [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 673.029893] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.055314] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd70349-d5b8-460a-b409-26dff8dcb362 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.059196] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Successfully created port: 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.064370] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08be75f8-1928-4111-87ca-5361afdb64c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.094208] env[62552]: DEBUG oslo_concurrency.lockutils [req-039d4772-b0b8-4e07-b0bb-987657a5a999 req-0dd01cea-4aa0-4789-8e58-3c4982db37e4 service nova] Releasing lock "refresh_cache-1b326e57-365e-49cc-b6f0-eed559ce32f5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.094562] env[62552]: DEBUG nova.compute.manager [req-039d4772-b0b8-4e07-b0bb-987657a5a999 req-0dd01cea-4aa0-4789-8e58-3c4982db37e4 service nova] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Received event network-vif-deleted-6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 673.095616] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb36cf32-f09b-4cba-a299-501afa200132 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.103044] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9916cf30-8351-422a-b730-1318d8aa3578 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.115995] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 673.118785] env[62552]: DEBUG nova.compute.provider_tree [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.628018] env[62552]: DEBUG nova.scheduler.client.report [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 674.031346] env[62552]: DEBUG nova.compute.manager [req-c496706c-454c-45d3-a2c9-70d6b90a31e2 req-426c007d-1073-4659-9f64-eeadae500c16 service nova] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Received event network-changed-9b274b33-c839-48c3-b8dd-92d5d4ccc0a8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 674.031682] env[62552]: DEBUG nova.compute.manager [req-c496706c-454c-45d3-a2c9-70d6b90a31e2 req-426c007d-1073-4659-9f64-eeadae500c16 service nova] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Refreshing instance network info cache due to event network-changed-9b274b33-c839-48c3-b8dd-92d5d4ccc0a8. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 674.031971] env[62552]: DEBUG oslo_concurrency.lockutils [req-c496706c-454c-45d3-a2c9-70d6b90a31e2 req-426c007d-1073-4659-9f64-eeadae500c16 service nova] Acquiring lock "refresh_cache-90b49142-4fd8-4c7b-98d7-0adaf5038c72" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.032378] env[62552]: DEBUG oslo_concurrency.lockutils [req-c496706c-454c-45d3-a2c9-70d6b90a31e2 req-426c007d-1073-4659-9f64-eeadae500c16 service nova] Acquired lock "refresh_cache-90b49142-4fd8-4c7b-98d7-0adaf5038c72" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.032627] env[62552]: DEBUG nova.network.neutron [req-c496706c-454c-45d3-a2c9-70d6b90a31e2 req-426c007d-1073-4659-9f64-eeadae500c16 service nova] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Refreshing network info cache for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 674.127116] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 674.134243] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.135078] env[62552]: ERROR nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48, please check neutron logs for more information. [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Traceback (most recent call last): [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self.driver.spawn(context, instance, image_meta, [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] vm_ref = self.build_virtual_machine(instance, [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.135078] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] for vif in network_info: [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] return self._sync_wrapper(fn, *args, **kwargs) [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self.wait() [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self[:] = self._gt.wait() [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] return self._exit_event.wait() [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] result = hub.switch() [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 674.135364] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] return self.greenlet.switch() [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] result = function(*args, **kwargs) [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] return func(*args, **kwargs) [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] raise e [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] nwinfo = self.network_api.allocate_for_instance( [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] created_port_ids = self._update_ports_for_instance( [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] with excutils.save_and_reraise_exception(): [ 674.135685] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] self.force_reraise() [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] raise self.value [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] updated_port = self._update_port( [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] _ensure_no_port_binding_failure(port) [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] raise exception.PortBindingFailed(port_id=port['id']) [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] nova.exception.PortBindingFailed: Binding failed for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48, please check neutron logs for more information. [ 674.136048] env[62552]: ERROR nova.compute.manager [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] [ 674.136310] env[62552]: DEBUG nova.compute.utils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Binding failed for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.138506] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.934s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.143324] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Build of instance b6bff07b-f316-4f26-9623-d50328df1ac0 was re-scheduled: Binding failed for port 1a3e3b55-d40a-4112-a72d-1d48e209ca48, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 674.144791] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 674.145208] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Acquiring lock "refresh_cache-b6bff07b-f316-4f26-9623-d50328df1ac0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.145411] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Acquired lock "refresh_cache-b6bff07b-f316-4f26-9623-d50328df1ac0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.145607] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.173508] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.174242] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.174467] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.178116] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.178116] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.178116] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.178116] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.178116] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.178288] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.178288] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.178288] env[62552]: DEBUG nova.virt.hardware [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.179991] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac36d1ba-7649-44b3-bb2d-9e2f3c8f7f34 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.193296] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7498c5ef-595d-4f05-8f6d-6bae6c8e6df1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.212300] env[62552]: ERROR nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8, please check neutron logs for more information. [ 674.212300] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 674.212300] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 674.212300] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 674.212300] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.212300] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 674.212300] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.212300] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 674.212300] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.212300] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 674.212300] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.212300] env[62552]: ERROR nova.compute.manager raise self.value [ 674.212300] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.212300] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 674.212300] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.212300] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 674.212888] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.212888] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 674.212888] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8, please check neutron logs for more information. [ 674.212888] env[62552]: ERROR nova.compute.manager [ 674.212888] env[62552]: Traceback (most recent call last): [ 674.212888] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 674.212888] env[62552]: listener.cb(fileno) [ 674.212888] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.212888] env[62552]: result = function(*args, **kwargs) [ 674.212888] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.212888] env[62552]: return func(*args, **kwargs) [ 674.212888] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 674.212888] env[62552]: raise e [ 674.212888] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 674.212888] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 674.212888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.212888] env[62552]: created_port_ids = self._update_ports_for_instance( [ 674.212888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.212888] env[62552]: with excutils.save_and_reraise_exception(): [ 674.212888] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.212888] env[62552]: self.force_reraise() [ 674.212888] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.212888] env[62552]: raise self.value [ 674.212888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.212888] env[62552]: updated_port = self._update_port( [ 674.212888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.212888] env[62552]: _ensure_no_port_binding_failure(port) [ 674.212888] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.212888] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 674.213631] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8, please check neutron logs for more information. [ 674.213631] env[62552]: Removing descriptor: 15 [ 674.213837] env[62552]: ERROR nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8, please check neutron logs for more information. [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Traceback (most recent call last): [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] yield resources [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self.driver.spawn(context, instance, image_meta, [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] vm_ref = self.build_virtual_machine(instance, [ 674.213837] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] for vif in network_info: [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] return self._sync_wrapper(fn, *args, **kwargs) [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self.wait() [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self[:] = self._gt.wait() [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] return self._exit_event.wait() [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 674.217218] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] result = hub.switch() [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] return self.greenlet.switch() [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] result = function(*args, **kwargs) [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] return func(*args, **kwargs) [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] raise e [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] nwinfo = self.network_api.allocate_for_instance( [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] created_port_ids = self._update_ports_for_instance( [ 674.217788] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] with excutils.save_and_reraise_exception(): [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self.force_reraise() [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] raise self.value [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] updated_port = self._update_port( [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] _ensure_no_port_binding_failure(port) [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] raise exception.PortBindingFailed(port_id=port['id']) [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] nova.exception.PortBindingFailed: Binding failed for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8, please check neutron logs for more information. [ 674.218168] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] [ 674.218531] env[62552]: INFO nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Terminating instance [ 674.551029] env[62552]: DEBUG nova.network.neutron [req-c496706c-454c-45d3-a2c9-70d6b90a31e2 req-426c007d-1073-4659-9f64-eeadae500c16 service nova] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.657953] env[62552]: DEBUG nova.network.neutron [req-c496706c-454c-45d3-a2c9-70d6b90a31e2 req-426c007d-1073-4659-9f64-eeadae500c16 service nova] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.675072] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.721939] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "refresh_cache-90b49142-4fd8-4c7b-98d7-0adaf5038c72" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.794635] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.154586] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410cd9fd-7f83-45e8-a03a-2180b3e70cda {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.162594] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524e9c11-7358-4646-bd28-30606764ba2b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.165989] env[62552]: DEBUG oslo_concurrency.lockutils [req-c496706c-454c-45d3-a2c9-70d6b90a31e2 req-426c007d-1073-4659-9f64-eeadae500c16 service nova] Releasing lock "refresh_cache-90b49142-4fd8-4c7b-98d7-0adaf5038c72" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.166456] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquired lock "refresh_cache-90b49142-4fd8-4c7b-98d7-0adaf5038c72" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.166647] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.196217] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a4cc3f-f198-434b-9cb9-c1307b411321 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.203987] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a1feae-9c28-4014-b4b7-758f12562c7d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.218527] env[62552]: DEBUG nova.compute.provider_tree [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.298023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Releasing lock "refresh_cache-b6bff07b-f316-4f26-9623-d50328df1ac0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.298274] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 675.298464] env[62552]: DEBUG nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 675.298633] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.312966] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.687440] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.721318] env[62552]: DEBUG nova.scheduler.client.report [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 675.799311] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.816275] env[62552]: DEBUG nova.network.neutron [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.055107] env[62552]: DEBUG nova.compute.manager [req-5567726b-8f67-4fe2-b29c-22ade22ef8e0 req-0a532f5b-3f94-4f8d-a786-5ee03f6c472d service nova] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Received event network-vif-deleted-9b274b33-c839-48c3-b8dd-92d5d4ccc0a8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 676.225810] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.087s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.226556] env[62552]: ERROR nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e, please check neutron logs for more information. [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Traceback (most recent call last): [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self.driver.spawn(context, instance, image_meta, [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] vm_ref = self.build_virtual_machine(instance, [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.226556] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] for vif in network_info: [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] return self._sync_wrapper(fn, *args, **kwargs) [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self.wait() [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self[:] = self._gt.wait() [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] return self._exit_event.wait() [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] result = hub.switch() [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 676.226892] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] return self.greenlet.switch() [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] result = function(*args, **kwargs) [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] return func(*args, **kwargs) [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] raise e [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] nwinfo = self.network_api.allocate_for_instance( [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] created_port_ids = self._update_ports_for_instance( [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] with excutils.save_and_reraise_exception(): [ 676.227451] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] self.force_reraise() [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] raise self.value [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] updated_port = self._update_port( [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] _ensure_no_port_binding_failure(port) [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] raise exception.PortBindingFailed(port_id=port['id']) [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] nova.exception.PortBindingFailed: Binding failed for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e, please check neutron logs for more information. [ 676.227809] env[62552]: ERROR nova.compute.manager [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] [ 676.228138] env[62552]: DEBUG nova.compute.utils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Binding failed for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.228597] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.733s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.228780] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.228879] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 676.229216] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.134s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.232030] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Build of instance 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7 was re-scheduled: Binding failed for port 4cbbecb5-081f-453a-9ce0-0462c6f1788e, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 676.232456] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 676.232745] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "refresh_cache-7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.232808] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "refresh_cache-7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.232963] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 676.234333] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af6c159-b5ce-4a82-8663-201c3467f33a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.243442] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e63dfe9-fc72-49fe-9772-e5f0d2f19eca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.258348] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b31141-848b-49d4-95d2-465071d2037c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.265043] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360a8965-ea23-41bd-bdba-7bcfdeeed7a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.297990] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181454MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 676.297990] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.301964] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Releasing lock "refresh_cache-90b49142-4fd8-4c7b-98d7-0adaf5038c72" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.302392] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 676.302586] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 676.303121] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e860756f-7abd-4b12-9dc0-87827d06717d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.310967] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890da8e7-fc10-468a-ae77-78a170c68aa8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.321897] env[62552]: INFO nova.compute.manager [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] [instance: b6bff07b-f316-4f26-9623-d50328df1ac0] Took 1.02 seconds to deallocate network for instance. [ 676.335806] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 90b49142-4fd8-4c7b-98d7-0adaf5038c72 could not be found. [ 676.336040] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.336230] env[62552]: INFO nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Took 0.03 seconds to destroy the instance on the hypervisor. [ 676.336467] env[62552]: DEBUG oslo.service.loopingcall [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.336698] env[62552]: DEBUG nova.compute.manager [-] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 676.336797] env[62552]: DEBUG nova.network.neutron [-] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.368826] env[62552]: DEBUG nova.network.neutron [-] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.757083] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.871797] env[62552]: DEBUG nova.network.neutron [-] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.880679] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.115895] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d2c23f-aeac-4848-afc6-12a21f109313 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.124815] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474a60ce-4e5b-46fe-a387-98632b091999 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.154621] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20762c7-cd97-427c-8fc1-0e6570b0adf9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.161836] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e24562e-6133-4168-a2be-7044aca8bba3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.174902] env[62552]: DEBUG nova.compute.provider_tree [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.354759] env[62552]: INFO nova.scheduler.client.report [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Deleted allocations for instance b6bff07b-f316-4f26-9623-d50328df1ac0 [ 677.377855] env[62552]: INFO nova.compute.manager [-] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Took 1.04 seconds to deallocate network for instance. [ 677.383393] env[62552]: DEBUG nova.compute.claims [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 677.383393] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.383393] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "refresh_cache-7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.384046] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 677.384046] env[62552]: DEBUG nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 677.384227] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 677.418659] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.677605] env[62552]: DEBUG nova.scheduler.client.report [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 677.868107] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0b775b87-7494-4119-90ef-6ba1da8193f1 tempest-ServerAddressesNegativeTestJSON-695296585 tempest-ServerAddressesNegativeTestJSON-695296585-project-member] Lock "b6bff07b-f316-4f26-9623-d50328df1ac0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.304s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.923032] env[62552]: DEBUG nova.network.neutron [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.187467] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.187467] env[62552]: ERROR nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 222de83b-1cfc-4f45-b375-c8ca83915c70, please check neutron logs for more information. [ 678.187467] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Traceback (most recent call last): [ 678.187467] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 678.187467] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self.driver.spawn(context, instance, image_meta, [ 678.187467] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 678.187467] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 678.187467] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 678.187467] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] vm_ref = self.build_virtual_machine(instance, [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] for vif in network_info: [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] return self._sync_wrapper(fn, *args, **kwargs) [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self.wait() [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self[:] = self._gt.wait() [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] return self._exit_event.wait() [ 678.187767] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] result = hub.switch() [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] return self.greenlet.switch() [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] result = function(*args, **kwargs) [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] return func(*args, **kwargs) [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] raise e [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] nwinfo = self.network_api.allocate_for_instance( [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.188147] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] created_port_ids = self._update_ports_for_instance( [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] with excutils.save_and_reraise_exception(): [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] self.force_reraise() [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] raise self.value [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] updated_port = self._update_port( [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] _ensure_no_port_binding_failure(port) [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.188462] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] raise exception.PortBindingFailed(port_id=port['id']) [ 678.188746] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] nova.exception.PortBindingFailed: Binding failed for port 222de83b-1cfc-4f45-b375-c8ca83915c70, please check neutron logs for more information. [ 678.188746] env[62552]: ERROR nova.compute.manager [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] [ 678.188746] env[62552]: DEBUG nova.compute.utils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Binding failed for port 222de83b-1cfc-4f45-b375-c8ca83915c70, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 678.190310] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Build of instance bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6 was re-scheduled: Binding failed for port 222de83b-1cfc-4f45-b375-c8ca83915c70, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 678.190475] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 678.192167] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "refresh_cache-bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.192167] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquired lock "refresh_cache-bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.192167] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.194043] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.810s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.371089] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 678.427525] env[62552]: INFO nova.compute.manager [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7] Took 1.04 seconds to deallocate network for instance. [ 678.719481] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.827221] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.899939] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.139779] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011d03aa-95f2-436d-8c12-a43801695739 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.147640] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba976bb-37b8-4cb8-aeed-3fcda9b20aa6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.177380] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96438a0a-0bd5-4191-b7e7-436efe84de46 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.184764] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8646b4cc-b038-46df-ae6a-f2c535cd7676 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.197952] env[62552]: DEBUG nova.compute.provider_tree [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.333803] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Releasing lock "refresh_cache-bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.334093] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 679.334283] env[62552]: DEBUG nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 679.334458] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.350157] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.475520] env[62552]: INFO nova.scheduler.client.report [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleted allocations for instance 7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7 [ 679.702654] env[62552]: DEBUG nova.scheduler.client.report [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 679.857737] env[62552]: DEBUG nova.network.neutron [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.983792] env[62552]: DEBUG oslo_concurrency.lockutils [None req-42ef3db2-486a-466c-b90c-ed85842c0b83 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "7d931c5c-7e2c-4ae9-a8b4-7fff3ed678f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.996s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.210258] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.211562] env[62552]: ERROR nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f9f76deb-cd3f-461a-a5ed-313f60107301, please check neutron logs for more information. [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Traceback (most recent call last): [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self.driver.spawn(context, instance, image_meta, [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] vm_ref = self.build_virtual_machine(instance, [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.211562] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] for vif in network_info: [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] return self._sync_wrapper(fn, *args, **kwargs) [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self.wait() [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self[:] = self._gt.wait() [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] return self._exit_event.wait() [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] result = hub.switch() [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.211888] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] return self.greenlet.switch() [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] result = function(*args, **kwargs) [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] return func(*args, **kwargs) [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] raise e [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] nwinfo = self.network_api.allocate_for_instance( [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] created_port_ids = self._update_ports_for_instance( [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] with excutils.save_and_reraise_exception(): [ 680.212209] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] self.force_reraise() [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] raise self.value [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] updated_port = self._update_port( [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] _ensure_no_port_binding_failure(port) [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] raise exception.PortBindingFailed(port_id=port['id']) [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] nova.exception.PortBindingFailed: Binding failed for port f9f76deb-cd3f-461a-a5ed-313f60107301, please check neutron logs for more information. [ 680.212517] env[62552]: ERROR nova.compute.manager [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] [ 680.212780] env[62552]: DEBUG nova.compute.utils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Binding failed for port f9f76deb-cd3f-461a-a5ed-313f60107301, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.214031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.777s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.217391] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Build of instance 74b58f3f-a575-4790-8a90-cad46b1499e6 was re-scheduled: Binding failed for port f9f76deb-cd3f-461a-a5ed-313f60107301, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 680.218127] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 680.218249] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-74b58f3f-a575-4790-8a90-cad46b1499e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.218456] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-74b58f3f-a575-4790-8a90-cad46b1499e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.218676] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.359764] env[62552]: INFO nova.compute.manager [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6] Took 1.03 seconds to deallocate network for instance. [ 680.488490] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 680.661784] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "714fb65d-9f80-4a81-a637-3e4398405d9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.661784] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "714fb65d-9f80-4a81-a637-3e4398405d9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.752610] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.912943] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.006591] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.182393] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7f6f8d-b699-403b-8283-9b2e9fd1a9be {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.190717] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53bf6ba7-b86a-4754-9ab2-a71c945e8263 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.227712] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6622d11f-0ab4-45ee-af3d-a2df5bafb2ac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.235455] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c4ef4d-fc35-47af-998c-d7350ef95d8e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.249406] env[62552]: DEBUG nova.compute.provider_tree [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.393381] env[62552]: INFO nova.scheduler.client.report [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Deleted allocations for instance bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6 [ 681.421015] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-74b58f3f-a575-4790-8a90-cad46b1499e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.421015] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 681.421015] env[62552]: DEBUG nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 681.421015] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.458388] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.752524] env[62552]: DEBUG nova.scheduler.client.report [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 681.823500] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "34736dd0-e617-475e-baa2-cb372db1afb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.823730] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "34736dd0-e617-475e-baa2-cb372db1afb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.903217] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b914421-a26f-4607-a853-c0776b0ffac0 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "bb8cf3bb-9e2a-44fe-a4c1-df39c52c79a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.804s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.959379] env[62552]: DEBUG nova.network.neutron [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.257795] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.044s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.258548] env[62552]: ERROR nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 98c5e091-4762-426f-84b0-68293670efc6, please check neutron logs for more information. [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Traceback (most recent call last): [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self.driver.spawn(context, instance, image_meta, [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] vm_ref = self.build_virtual_machine(instance, [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.258548] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] for vif in network_info: [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] return self._sync_wrapper(fn, *args, **kwargs) [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self.wait() [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self[:] = self._gt.wait() [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] return self._exit_event.wait() [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] result = hub.switch() [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 682.258857] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] return self.greenlet.switch() [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] result = function(*args, **kwargs) [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] return func(*args, **kwargs) [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] raise e [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] nwinfo = self.network_api.allocate_for_instance( [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] created_port_ids = self._update_ports_for_instance( [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] with excutils.save_and_reraise_exception(): [ 682.259210] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] self.force_reraise() [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] raise self.value [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] updated_port = self._update_port( [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] _ensure_no_port_binding_failure(port) [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] raise exception.PortBindingFailed(port_id=port['id']) [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] nova.exception.PortBindingFailed: Binding failed for port 98c5e091-4762-426f-84b0-68293670efc6, please check neutron logs for more information. [ 682.259533] env[62552]: ERROR nova.compute.manager [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] [ 682.259804] env[62552]: DEBUG nova.compute.utils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Binding failed for port 98c5e091-4762-426f-84b0-68293670efc6, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.260566] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.895s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.267019] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Build of instance 384445d0-93b0-427c-92aa-ef557351a7cd was re-scheduled: Binding failed for port 98c5e091-4762-426f-84b0-68293670efc6, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 682.267019] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 682.267019] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Acquiring lock "refresh_cache-384445d0-93b0-427c-92aa-ef557351a7cd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.267019] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Acquired lock "refresh_cache-384445d0-93b0-427c-92aa-ef557351a7cd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.267261] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.409566] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 682.465771] env[62552]: INFO nova.compute.manager [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 74b58f3f-a575-4790-8a90-cad46b1499e6] Took 1.05 seconds to deallocate network for instance. [ 682.793981] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.870095] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.932094] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.187485] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218b99cc-a141-4803-a7e5-f256b416e4ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.195241] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3661142a-9f86-4732-b7bb-108a08411f40 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.225290] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab51e6e-1e8b-4b22-a033-f4c6c32618cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.233244] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb38230-d2c9-48b1-bf0c-4f0acdb9fcde {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.246633] env[62552]: DEBUG nova.compute.provider_tree [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.373246] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Releasing lock "refresh_cache-384445d0-93b0-427c-92aa-ef557351a7cd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.373860] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 683.373860] env[62552]: DEBUG nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 683.373860] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.409761] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.505411] env[62552]: INFO nova.scheduler.client.report [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted allocations for instance 74b58f3f-a575-4790-8a90-cad46b1499e6 [ 683.749371] env[62552]: DEBUG nova.scheduler.client.report [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 683.915135] env[62552]: DEBUG nova.network.neutron [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.014830] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e1b8bb78-c287-4ef2-b51e-97e15c0b2c0c tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "74b58f3f-a575-4790-8a90-cad46b1499e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.354s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.254322] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.254721] env[62552]: ERROR nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723, please check neutron logs for more information. [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Traceback (most recent call last): [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self.driver.spawn(context, instance, image_meta, [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] vm_ref = self.build_virtual_machine(instance, [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.254721] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] for vif in network_info: [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] return self._sync_wrapper(fn, *args, **kwargs) [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self.wait() [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self[:] = self._gt.wait() [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] return self._exit_event.wait() [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] result = hub.switch() [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.255029] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] return self.greenlet.switch() [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] result = function(*args, **kwargs) [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] return func(*args, **kwargs) [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] raise e [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] nwinfo = self.network_api.allocate_for_instance( [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] created_port_ids = self._update_ports_for_instance( [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] with excutils.save_and_reraise_exception(): [ 684.255345] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] self.force_reraise() [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] raise self.value [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] updated_port = self._update_port( [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] _ensure_no_port_binding_failure(port) [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] raise exception.PortBindingFailed(port_id=port['id']) [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] nova.exception.PortBindingFailed: Binding failed for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723, please check neutron logs for more information. [ 684.255653] env[62552]: ERROR nova.compute.manager [instance: fae607f4-e18a-4cd4-8843-246a3341a548] [ 684.255916] env[62552]: DEBUG nova.compute.utils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Binding failed for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.256908] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.072s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.258375] env[62552]: INFO nova.compute.claims [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.261301] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Build of instance fae607f4-e18a-4cd4-8843-246a3341a548 was re-scheduled: Binding failed for port 700526b4-3e5f-4f5f-8f32-0b6cb2778723, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 684.261912] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 684.262183] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquiring lock "refresh_cache-fae607f4-e18a-4cd4-8843-246a3341a548" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.262363] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Acquired lock "refresh_cache-fae607f4-e18a-4cd4-8843-246a3341a548" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.262569] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 684.418953] env[62552]: INFO nova.compute.manager [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] [instance: 384445d0-93b0-427c-92aa-ef557351a7cd] Took 1.04 seconds to deallocate network for instance. [ 684.518155] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 684.757332] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "36233c83-1170-4a7a-be0b-6a0e8b139de7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.758417] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.789524] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.876561] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.040255] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.378505] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Releasing lock "refresh_cache-fae607f4-e18a-4cd4-8843-246a3341a548" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.378738] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 685.378903] env[62552]: DEBUG nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 685.379081] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.399216] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.449164] env[62552]: INFO nova.scheduler.client.report [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Deleted allocations for instance 384445d0-93b0-427c-92aa-ef557351a7cd [ 685.629142] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929379fa-1293-430a-841b-8083f1b672ea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.636956] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f6960e-620e-4d9e-bd3c-805cd33e564c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.669185] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07fd551-ad28-4b5c-a25a-7e9cc769cbcb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.674464] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "3b333977-0b4b-4ccf-bed5-231215abae3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.674706] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.680165] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5db9557-ebd2-4311-b059-6cbdaba39c54 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.693965] env[62552]: DEBUG nova.compute.provider_tree [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.902226] env[62552]: DEBUG nova.network.neutron [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.958844] env[62552]: DEBUG oslo_concurrency.lockutils [None req-720fa749-6469-494f-93fc-44269356a46a tempest-ServersAdminNegativeTestJSON-1404624803 tempest-ServersAdminNegativeTestJSON-1404624803-project-member] Lock "384445d0-93b0-427c-92aa-ef557351a7cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.912s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.197111] env[62552]: DEBUG nova.scheduler.client.report [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 686.404378] env[62552]: INFO nova.compute.manager [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] [instance: fae607f4-e18a-4cd4-8843-246a3341a548] Took 1.03 seconds to deallocate network for instance. [ 686.461791] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 686.704654] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.705378] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 686.711020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.882s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.984882] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.215195] env[62552]: DEBUG nova.compute.utils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.219472] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 687.219639] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 687.288841] env[62552]: DEBUG nova.policy [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '81ff146ac33b4a5ab1b91df0614c1cfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '173a8b2a661143bb947461a872b418e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.442175] env[62552]: INFO nova.scheduler.client.report [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Deleted allocations for instance fae607f4-e18a-4cd4-8843-246a3341a548 [ 687.637383] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f74742-8348-4840-bf53-a5d0d6187e98 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.645115] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e08662-9d45-45d6-a56f-b51ad2d36bab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.679384] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bea49e3-7ab8-4b39-8403-961928a4f5da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.686037] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Successfully created port: 8323852d-b171-4826-ba44-55a07704e2b1 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.690354] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4290fb99-ef51-4390-84cc-99c215345145 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.708247] env[62552]: DEBUG nova.compute.provider_tree [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.720060] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 687.953706] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95a6fcf3-8855-495c-a4ec-b7312e7d27b7 tempest-MigrationsAdminTest-964455516 tempest-MigrationsAdminTest-964455516-project-member] Lock "fae607f4-e18a-4cd4-8843-246a3341a548" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.452s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.212493] env[62552]: DEBUG nova.scheduler.client.report [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 688.458296] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 688.718991] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.010s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.719868] env[62552]: ERROR nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 43393080-f145-4d14-8609-4c60eb83b25a, please check neutron logs for more information. [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Traceback (most recent call last): [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self.driver.spawn(context, instance, image_meta, [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] vm_ref = self.build_virtual_machine(instance, [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.719868] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] for vif in network_info: [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] return self._sync_wrapper(fn, *args, **kwargs) [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self.wait() [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self[:] = self._gt.wait() [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] return self._exit_event.wait() [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] result = hub.switch() [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 688.720228] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] return self.greenlet.switch() [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] result = function(*args, **kwargs) [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] return func(*args, **kwargs) [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] raise e [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] nwinfo = self.network_api.allocate_for_instance( [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] created_port_ids = self._update_ports_for_instance( [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] with excutils.save_and_reraise_exception(): [ 688.720542] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] self.force_reraise() [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] raise self.value [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] updated_port = self._update_port( [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] _ensure_no_port_binding_failure(port) [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] raise exception.PortBindingFailed(port_id=port['id']) [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] nova.exception.PortBindingFailed: Binding failed for port 43393080-f145-4d14-8609-4c60eb83b25a, please check neutron logs for more information. [ 688.720858] env[62552]: ERROR nova.compute.manager [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] [ 688.721138] env[62552]: DEBUG nova.compute.utils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Binding failed for port 43393080-f145-4d14-8609-4c60eb83b25a, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 688.722877] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.693s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.727705] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Build of instance 187dd75d-cdb0-4b86-a396-f43778fdcc21 was re-scheduled: Binding failed for port 43393080-f145-4d14-8609-4c60eb83b25a, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 688.728255] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 688.728733] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "refresh_cache-187dd75d-cdb0-4b86-a396-f43778fdcc21" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.728950] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquired lock "refresh_cache-187dd75d-cdb0-4b86-a396-f43778fdcc21" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.729176] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.738221] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 688.778173] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.778506] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.778708] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.779590] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.780184] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.780521] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.780933] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.781153] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.781377] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.781629] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.781984] env[62552]: DEBUG nova.virt.hardware [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.784368] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0a4deb-fa9b-46a8-ad78-6779e75cda50 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.794118] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7b8eed-55c7-4e98-8b81-0b1e3f0fb1e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.799409] env[62552]: DEBUG nova.compute.manager [req-bbf79678-988c-4c05-af71-78a17724cbbc req-12f632ad-0d12-444b-9dd9-bb6d0abfa555 service nova] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Received event network-changed-8323852d-b171-4826-ba44-55a07704e2b1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 688.799787] env[62552]: DEBUG nova.compute.manager [req-bbf79678-988c-4c05-af71-78a17724cbbc req-12f632ad-0d12-444b-9dd9-bb6d0abfa555 service nova] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Refreshing instance network info cache due to event network-changed-8323852d-b171-4826-ba44-55a07704e2b1. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 688.800384] env[62552]: DEBUG oslo_concurrency.lockutils [req-bbf79678-988c-4c05-af71-78a17724cbbc req-12f632ad-0d12-444b-9dd9-bb6d0abfa555 service nova] Acquiring lock "refresh_cache-dd4c1e35-0397-4ee0-932d-2dc7f0870f84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.800679] env[62552]: DEBUG oslo_concurrency.lockutils [req-bbf79678-988c-4c05-af71-78a17724cbbc req-12f632ad-0d12-444b-9dd9-bb6d0abfa555 service nova] Acquired lock "refresh_cache-dd4c1e35-0397-4ee0-932d-2dc7f0870f84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.801224] env[62552]: DEBUG nova.network.neutron [req-bbf79678-988c-4c05-af71-78a17724cbbc req-12f632ad-0d12-444b-9dd9-bb6d0abfa555 service nova] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Refreshing network info cache for port 8323852d-b171-4826-ba44-55a07704e2b1 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 688.982745] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.078374] env[62552]: ERROR nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8323852d-b171-4826-ba44-55a07704e2b1, please check neutron logs for more information. [ 689.078374] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.078374] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 689.078374] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.078374] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.078374] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.078374] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.078374] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.078374] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.078374] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 689.078374] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.078374] env[62552]: ERROR nova.compute.manager raise self.value [ 689.078374] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.078374] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.078374] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.078374] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.078777] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.078777] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.078777] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8323852d-b171-4826-ba44-55a07704e2b1, please check neutron logs for more information. [ 689.078777] env[62552]: ERROR nova.compute.manager [ 689.078777] env[62552]: Traceback (most recent call last): [ 689.078777] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.078777] env[62552]: listener.cb(fileno) [ 689.078777] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.078777] env[62552]: result = function(*args, **kwargs) [ 689.078777] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.078777] env[62552]: return func(*args, **kwargs) [ 689.078777] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 689.078777] env[62552]: raise e [ 689.078777] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 689.078777] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 689.078777] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.078777] env[62552]: created_port_ids = self._update_ports_for_instance( [ 689.078777] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.078777] env[62552]: with excutils.save_and_reraise_exception(): [ 689.078777] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.078777] env[62552]: self.force_reraise() [ 689.078777] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.078777] env[62552]: raise self.value [ 689.078777] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.078777] env[62552]: updated_port = self._update_port( [ 689.078777] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.078777] env[62552]: _ensure_no_port_binding_failure(port) [ 689.078777] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.078777] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.079578] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 8323852d-b171-4826-ba44-55a07704e2b1, please check neutron logs for more information. [ 689.079578] env[62552]: Removing descriptor: 16 [ 689.079578] env[62552]: ERROR nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8323852d-b171-4826-ba44-55a07704e2b1, please check neutron logs for more information. [ 689.079578] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Traceback (most recent call last): [ 689.079578] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 689.079578] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] yield resources [ 689.079578] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 689.079578] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self.driver.spawn(context, instance, image_meta, [ 689.079578] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 689.079578] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.079578] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.079578] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] vm_ref = self.build_virtual_machine(instance, [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] for vif in network_info: [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] return self._sync_wrapper(fn, *args, **kwargs) [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self.wait() [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self[:] = self._gt.wait() [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] return self._exit_event.wait() [ 689.079934] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] result = hub.switch() [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] return self.greenlet.switch() [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] result = function(*args, **kwargs) [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] return func(*args, **kwargs) [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] raise e [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] nwinfo = self.network_api.allocate_for_instance( [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.080309] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] created_port_ids = self._update_ports_for_instance( [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] with excutils.save_and_reraise_exception(): [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self.force_reraise() [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] raise self.value [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] updated_port = self._update_port( [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] _ensure_no_port_binding_failure(port) [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.080654] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] raise exception.PortBindingFailed(port_id=port['id']) [ 689.080969] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] nova.exception.PortBindingFailed: Binding failed for port 8323852d-b171-4826-ba44-55a07704e2b1, please check neutron logs for more information. [ 689.080969] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] [ 689.080969] env[62552]: INFO nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Terminating instance [ 689.262060] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.339350] env[62552]: DEBUG nova.network.neutron [req-bbf79678-988c-4c05-af71-78a17724cbbc req-12f632ad-0d12-444b-9dd9-bb6d0abfa555 service nova] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.443266] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.484429] env[62552]: DEBUG nova.network.neutron [req-bbf79678-988c-4c05-af71-78a17724cbbc req-12f632ad-0d12-444b-9dd9-bb6d0abfa555 service nova] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.587535] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "refresh_cache-dd4c1e35-0397-4ee0-932d-2dc7f0870f84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.693510] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81319be6-801c-4c6d-89ed-f093533074bb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.705635] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc14b06-b23f-4d81-a2ef-6c5608efef85 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.743281] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4adf37bc-a2bc-4658-8358-6e3cc946a0d5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.751439] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c800d7f-55e2-4b4e-963d-26800132e2fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.765783] env[62552]: DEBUG nova.compute.provider_tree [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.949024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Releasing lock "refresh_cache-187dd75d-cdb0-4b86-a396-f43778fdcc21" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.949024] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 689.949024] env[62552]: DEBUG nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 689.949024] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.979302] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.987718] env[62552]: DEBUG oslo_concurrency.lockutils [req-bbf79678-988c-4c05-af71-78a17724cbbc req-12f632ad-0d12-444b-9dd9-bb6d0abfa555 service nova] Releasing lock "refresh_cache-dd4c1e35-0397-4ee0-932d-2dc7f0870f84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.988161] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquired lock "refresh_cache-dd4c1e35-0397-4ee0-932d-2dc7f0870f84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.988409] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.122384] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquiring lock "fef94e30-f946-4d7f-a108-f32cad505bc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.122524] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "fef94e30-f946-4d7f-a108-f32cad505bc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.274640] env[62552]: DEBUG nova.scheduler.client.report [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 690.484675] env[62552]: DEBUG nova.network.neutron [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.510445] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.777598] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.054s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.779441] env[62552]: ERROR nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8, please check neutron logs for more information. [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Traceback (most recent call last): [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self.driver.spawn(context, instance, image_meta, [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] vm_ref = self.build_virtual_machine(instance, [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.779441] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] for vif in network_info: [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] return self._sync_wrapper(fn, *args, **kwargs) [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self.wait() [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self[:] = self._gt.wait() [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] return self._exit_event.wait() [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] result = hub.switch() [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 690.780078] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] return self.greenlet.switch() [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] result = function(*args, **kwargs) [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] return func(*args, **kwargs) [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] raise e [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] nwinfo = self.network_api.allocate_for_instance( [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] created_port_ids = self._update_ports_for_instance( [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] with excutils.save_and_reraise_exception(): [ 690.780773] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] self.force_reraise() [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] raise self.value [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] updated_port = self._update_port( [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] _ensure_no_port_binding_failure(port) [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] raise exception.PortBindingFailed(port_id=port['id']) [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] nova.exception.PortBindingFailed: Binding failed for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8, please check neutron logs for more information. [ 690.781102] env[62552]: ERROR nova.compute.manager [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] [ 690.781376] env[62552]: DEBUG nova.compute.utils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Binding failed for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 690.782702] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.485s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.785095] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Build of instance 1b326e57-365e-49cc-b6f0-eed559ce32f5 was re-scheduled: Binding failed for port 6bb82b44-a1b3-4827-8c1a-1bb2fb1bfde8, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 690.785569] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 690.785753] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquiring lock "refresh_cache-1b326e57-365e-49cc-b6f0-eed559ce32f5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.785899] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Acquired lock "refresh_cache-1b326e57-365e-49cc-b6f0-eed559ce32f5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.786704] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.792984] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.988265] env[62552]: INFO nova.compute.manager [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: 187dd75d-cdb0-4b86-a396-f43778fdcc21] Took 1.04 seconds to deallocate network for instance. [ 691.061334] env[62552]: DEBUG nova.compute.manager [req-9fb40385-0fa4-4e48-b254-f1d665d4456e req-2562935c-0450-423d-944f-fe6c2c5502fa service nova] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Received event network-vif-deleted-8323852d-b171-4826-ba44-55a07704e2b1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 691.298355] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Releasing lock "refresh_cache-dd4c1e35-0397-4ee0-932d-2dc7f0870f84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.302020] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 691.302020] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 691.302020] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b56232b-a785-4357-a86e-f949bf64ec2d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.311223] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a25e7c-998f-475c-bea1-b172579e1825 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.328419] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.345769] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dd4c1e35-0397-4ee0-932d-2dc7f0870f84 could not be found. [ 691.345769] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.345769] env[62552]: INFO nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Took 0.05 seconds to destroy the instance on the hypervisor. [ 691.345769] env[62552]: DEBUG oslo.service.loopingcall [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.345769] env[62552]: DEBUG nova.compute.manager [-] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 691.345769] env[62552]: DEBUG nova.network.neutron [-] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.377105] env[62552]: DEBUG nova.network.neutron [-] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.431673] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.878723] env[62552]: DEBUG nova.network.neutron [-] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.935664] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Releasing lock "refresh_cache-1b326e57-365e-49cc-b6f0-eed559ce32f5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.935919] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 691.936106] env[62552]: DEBUG nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 691.936279] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.957508] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.023550] env[62552]: INFO nova.scheduler.client.report [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Deleted allocations for instance 187dd75d-cdb0-4b86-a396-f43778fdcc21 [ 692.333019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 187dd75d-cdb0-4b86-a396-f43778fdcc21 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.383081] env[62552]: INFO nova.compute.manager [-] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Took 1.04 seconds to deallocate network for instance. [ 692.384202] env[62552]: DEBUG nova.compute.claims [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 692.384202] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.460625] env[62552]: DEBUG nova.network.neutron [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.534105] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ce9bc28-d113-41db-8451-6bbc49d7b894 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "187dd75d-cdb0-4b86-a396-f43778fdcc21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.367s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.835360] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 1b326e57-365e-49cc-b6f0-eed559ce32f5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.835360] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 90b49142-4fd8-4c7b-98d7-0adaf5038c72 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 692.835360] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance dd4c1e35-0397-4ee0-932d-2dc7f0870f84 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 692.966470] env[62552]: INFO nova.compute.manager [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] [instance: 1b326e57-365e-49cc-b6f0-eed559ce32f5] Took 1.03 seconds to deallocate network for instance. [ 693.038980] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 693.337059] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.559995] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.939716] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.002155] env[62552]: INFO nova.scheduler.client.report [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Deleted allocations for instance 1b326e57-365e-49cc-b6f0-eed559ce32f5 [ 694.351760] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 04289f4b-4342-4377-8f9c-3013eb646cc6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.510532] env[62552]: DEBUG oslo_concurrency.lockutils [None req-26571a3b-db6b-4240-89b1-478a9a04df0f tempest-DeleteServersAdminTestJSON-1504139255 tempest-DeleteServersAdminTestJSON-1504139255-project-member] Lock "1b326e57-365e-49cc-b6f0-eed559ce32f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.911s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.854644] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance f8f37b3c-5e74-4459-9177-aacfac707616 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.013484] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 695.360239] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 2031fc3f-2c79-4f6e-b7eb-29235814141d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.535540] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.862370] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance cf0dcba7-a36e-435c-8c34-d352b49a723b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.177189] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "80478878-ff82-4ed6-a851-8eb2bec01e22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.177502] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.365412] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance a9a16b57-603a-4d43-b834-e6ea62675f3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.868833] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 5ae752f6-9fd4-46fd-a7d2-558af22e2943 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.372922] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance cfdee9ce-8c90-40cc-a09e-2386c719c02e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.877272] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance e8ec980c-0186-48cb-84b9-af902f23d2fb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.984302] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.984545] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.380526] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance ab363b52-ade2-45c4-ba3a-ae46ffaf831e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.883982] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 1c7e408a-392d-4fcf-bae8-588ed9f92fec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.387229] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 1252f45a-d55d-44c0-8629-aba51fb89956 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.891140] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance cdee22a9-4327-47af-ab14-8403a06b2802 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.394609] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 1472ca01-ba04-4e7d-a9cf-ccaf29c04282 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.897890] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 8707ce4b-677e-4f13-86f8-3e327d19380b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.402289] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance ce5d0165-65f1-4505-9c46-1129c56a8913 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.906009] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 0418260a-aa27-4955-ab15-b180ec04f0b3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 702.409959] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 00b06df7-e7ae-48fd-8887-65647931d5cf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 702.912165] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 4b80bc41-1b63-444e-8039-696e3d78d9a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.382785] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.383030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.414648] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 51bbcc4e-8251-4b38-9d36-8aea8fc7705d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.917631] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 714fb65d-9f80-4a81-a637-3e4398405d9b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.420145] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 34736dd0-e617-475e-baa2-cb372db1afb2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.924730] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 36233c83-1170-4a7a-be0b-6a0e8b139de7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.427380] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 3b333977-0b4b-4ccf-bed5-231215abae3c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.929861] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance fef94e30-f946-4d7f-a108-f32cad505bc4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.930154] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 705.930309] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 706.289666] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c233bf07-2038-435c-98c0-ebf3ccfc3544 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.297257] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7355960-7220-4d57-8cd0-fe28e571c28f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.328715] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9578b6-0907-4823-ab5a-a6eb5849e1a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.336071] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff994ee-f85f-4757-b895-c878a368f26e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.349895] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.852124] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 707.356053] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 707.356167] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 16.573s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.356416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.974s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.164707] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d43f582-7619-4058-a1cf-24cc6d7b2500 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.172364] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf3dcc5-d1ee-4d29-a226-7b689f26f4db {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.203113] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67ff78a-c481-4414-84ac-3bb7ff8380c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.210291] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34cf1f0-0cea-4260-89ca-e23aa8b65184 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.223555] env[62552]: DEBUG nova.compute.provider_tree [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.730888] env[62552]: DEBUG nova.scheduler.client.report [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 709.235940] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.879s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.237892] env[62552]: ERROR nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8, please check neutron logs for more information. [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Traceback (most recent call last): [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self.driver.spawn(context, instance, image_meta, [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] vm_ref = self.build_virtual_machine(instance, [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.237892] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] for vif in network_info: [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] return self._sync_wrapper(fn, *args, **kwargs) [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self.wait() [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self[:] = self._gt.wait() [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] return self._exit_event.wait() [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] result = hub.switch() [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 709.238294] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] return self.greenlet.switch() [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] result = function(*args, **kwargs) [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] return func(*args, **kwargs) [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] raise e [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] nwinfo = self.network_api.allocate_for_instance( [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] created_port_ids = self._update_ports_for_instance( [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] with excutils.save_and_reraise_exception(): [ 709.238688] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] self.force_reraise() [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] raise self.value [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] updated_port = self._update_port( [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] _ensure_no_port_binding_failure(port) [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] raise exception.PortBindingFailed(port_id=port['id']) [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] nova.exception.PortBindingFailed: Binding failed for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8, please check neutron logs for more information. [ 709.239079] env[62552]: ERROR nova.compute.manager [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] [ 709.239547] env[62552]: DEBUG nova.compute.utils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Binding failed for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 709.239547] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.339s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.240018] env[62552]: INFO nova.compute.claims [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.242602] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Build of instance 90b49142-4fd8-4c7b-98d7-0adaf5038c72 was re-scheduled: Binding failed for port 9b274b33-c839-48c3-b8dd-92d5d4ccc0a8, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 709.243043] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 709.243268] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "refresh_cache-90b49142-4fd8-4c7b-98d7-0adaf5038c72" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.243414] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquired lock "refresh_cache-90b49142-4fd8-4c7b-98d7-0adaf5038c72" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.243570] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.766565] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.847735] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.350937] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Releasing lock "refresh_cache-90b49142-4fd8-4c7b-98d7-0adaf5038c72" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.350937] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 710.350937] env[62552]: DEBUG nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 710.350937] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.379854] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.598172] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a982da6-bdc3-48eb-beff-5a154e86083e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.606032] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09473465-3c4a-4b23-8fc6-fc5c33fcd867 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.636927] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b8258a-4e1a-4903-ad7f-aa817b667fb9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.644412] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c187099-afcf-4131-959e-104fb8a4d70c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.657869] env[62552]: DEBUG nova.compute.provider_tree [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.882703] env[62552]: DEBUG nova.network.neutron [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.161321] env[62552]: DEBUG nova.scheduler.client.report [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 711.385704] env[62552]: INFO nova.compute.manager [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 90b49142-4fd8-4c7b-98d7-0adaf5038c72] Took 1.03 seconds to deallocate network for instance. [ 711.665678] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.666241] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 711.668910] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.662s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.670401] env[62552]: INFO nova.compute.claims [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.176128] env[62552]: DEBUG nova.compute.utils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.177544] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 712.177717] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 712.235559] env[62552]: DEBUG nova.policy [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8084fec9987f406e80126a0f684f09f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd78973494f0c46d1a34d0392329b1e79', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.414748] env[62552]: INFO nova.scheduler.client.report [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Deleted allocations for instance 90b49142-4fd8-4c7b-98d7-0adaf5038c72 [ 712.622253] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Successfully created port: 7221d21a-c271-4bb8-b1e9-821a44f22b4f {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.682201] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 712.924335] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1472ff82-3c30-4303-8514-c67090542fb9 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "90b49142-4fd8-4c7b-98d7-0adaf5038c72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.960s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.024662] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45b56e4-bdb9-4f90-9049-f5ec99e2da0a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.032628] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e7c420-0c26-4488-b191-f37af3b4e905 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.064488] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84c0531-5162-435d-a8cb-c469aba52ec0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.071954] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2977873c-c0c1-469f-b8f1-85aee5272ce7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.088578] env[62552]: DEBUG nova.compute.provider_tree [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.427967] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 713.576982] env[62552]: DEBUG nova.compute.manager [req-4679f2cc-adb6-4b1d-b0d4-376c29d0a133 req-38a31be2-8960-4b0d-afba-ce260356895c service nova] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Received event network-changed-7221d21a-c271-4bb8-b1e9-821a44f22b4f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 713.577199] env[62552]: DEBUG nova.compute.manager [req-4679f2cc-adb6-4b1d-b0d4-376c29d0a133 req-38a31be2-8960-4b0d-afba-ce260356895c service nova] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Refreshing instance network info cache due to event network-changed-7221d21a-c271-4bb8-b1e9-821a44f22b4f. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 713.577413] env[62552]: DEBUG oslo_concurrency.lockutils [req-4679f2cc-adb6-4b1d-b0d4-376c29d0a133 req-38a31be2-8960-4b0d-afba-ce260356895c service nova] Acquiring lock "refresh_cache-4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.577595] env[62552]: DEBUG oslo_concurrency.lockutils [req-4679f2cc-adb6-4b1d-b0d4-376c29d0a133 req-38a31be2-8960-4b0d-afba-ce260356895c service nova] Acquired lock "refresh_cache-4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.577781] env[62552]: DEBUG nova.network.neutron [req-4679f2cc-adb6-4b1d-b0d4-376c29d0a133 req-38a31be2-8960-4b0d-afba-ce260356895c service nova] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Refreshing network info cache for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.591365] env[62552]: DEBUG nova.scheduler.client.report [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 713.690771] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 713.716806] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 713.717020] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 713.717177] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 713.717394] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 713.717500] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 713.717664] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 713.718246] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 713.718246] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 713.718246] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 713.718431] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 713.718626] env[62552]: DEBUG nova.virt.hardware [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 713.719573] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d996f25-6b3c-4828-975c-76d777a8425f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.727879] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8cef6e-64e3-4c5e-88c4-d10b24ee49e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.787295] env[62552]: ERROR nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f, please check neutron logs for more information. [ 713.787295] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 713.787295] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 713.787295] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 713.787295] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.787295] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 713.787295] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.787295] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 713.787295] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.787295] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 713.787295] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.787295] env[62552]: ERROR nova.compute.manager raise self.value [ 713.787295] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.787295] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 713.787295] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.787295] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 713.787828] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.787828] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 713.787828] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f, please check neutron logs for more information. [ 713.787828] env[62552]: ERROR nova.compute.manager [ 713.787828] env[62552]: Traceback (most recent call last): [ 713.787828] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 713.787828] env[62552]: listener.cb(fileno) [ 713.787828] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.787828] env[62552]: result = function(*args, **kwargs) [ 713.787828] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.787828] env[62552]: return func(*args, **kwargs) [ 713.787828] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 713.787828] env[62552]: raise e [ 713.787828] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 713.787828] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 713.787828] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.787828] env[62552]: created_port_ids = self._update_ports_for_instance( [ 713.787828] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.787828] env[62552]: with excutils.save_and_reraise_exception(): [ 713.787828] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.787828] env[62552]: self.force_reraise() [ 713.787828] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.787828] env[62552]: raise self.value [ 713.787828] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.787828] env[62552]: updated_port = self._update_port( [ 713.787828] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.787828] env[62552]: _ensure_no_port_binding_failure(port) [ 713.787828] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.787828] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 713.788705] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f, please check neutron logs for more information. [ 713.788705] env[62552]: Removing descriptor: 15 [ 713.788705] env[62552]: ERROR nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f, please check neutron logs for more information. [ 713.788705] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Traceback (most recent call last): [ 713.788705] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 713.788705] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] yield resources [ 713.788705] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 713.788705] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self.driver.spawn(context, instance, image_meta, [ 713.788705] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 713.788705] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.788705] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.788705] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] vm_ref = self.build_virtual_machine(instance, [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] for vif in network_info: [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] return self._sync_wrapper(fn, *args, **kwargs) [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self.wait() [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self[:] = self._gt.wait() [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] return self._exit_event.wait() [ 713.789088] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] result = hub.switch() [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] return self.greenlet.switch() [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] result = function(*args, **kwargs) [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] return func(*args, **kwargs) [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] raise e [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] nwinfo = self.network_api.allocate_for_instance( [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.789498] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] created_port_ids = self._update_ports_for_instance( [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] with excutils.save_and_reraise_exception(): [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self.force_reraise() [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] raise self.value [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] updated_port = self._update_port( [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] _ensure_no_port_binding_failure(port) [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.789892] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] raise exception.PortBindingFailed(port_id=port['id']) [ 713.790266] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] nova.exception.PortBindingFailed: Binding failed for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f, please check neutron logs for more information. [ 713.790266] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] [ 713.790266] env[62552]: INFO nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Terminating instance [ 713.954993] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.095979] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.096500] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 714.099070] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.167s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.100515] env[62552]: INFO nova.compute.claims [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.104599] env[62552]: DEBUG nova.network.neutron [req-4679f2cc-adb6-4b1d-b0d4-376c29d0a133 req-38a31be2-8960-4b0d-afba-ce260356895c service nova] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.216418] env[62552]: DEBUG nova.network.neutron [req-4679f2cc-adb6-4b1d-b0d4-376c29d0a133 req-38a31be2-8960-4b0d-afba-ce260356895c service nova] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.292977] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Acquiring lock "refresh_cache-4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.607957] env[62552]: DEBUG nova.compute.utils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 714.609355] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 714.609580] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 714.657027] env[62552]: DEBUG nova.policy [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '81ff146ac33b4a5ab1b91df0614c1cfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '173a8b2a661143bb947461a872b418e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 714.720618] env[62552]: DEBUG oslo_concurrency.lockutils [req-4679f2cc-adb6-4b1d-b0d4-376c29d0a133 req-38a31be2-8960-4b0d-afba-ce260356895c service nova] Releasing lock "refresh_cache-4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.720618] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Acquired lock "refresh_cache-4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.720618] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.963821] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Successfully created port: 4cc71534-52d9-4238-8ab9-8384cf3ed303 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.113756] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 715.261076] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.450028] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.484026] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95748674-79f9-4089-b2f3-796c00415501 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.489648] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b895589-51a2-4ae1-bd1c-0adc8bcdc460 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.522305] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18ffa17-b6da-4a8f-82f4-0e568f64ed33 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.529548] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9f1a77-fa42-435e-bdc2-a8fea0b416f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.542369] env[62552]: DEBUG nova.compute.provider_tree [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.683243] env[62552]: DEBUG nova.compute.manager [req-14ec38e5-6605-4a10-bce8-fdf3f7549916 req-63001e3c-ac79-4d0f-b6a0-e227989106f3 service nova] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Received event network-vif-deleted-7221d21a-c271-4bb8-b1e9-821a44f22b4f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 715.953758] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Releasing lock "refresh_cache-4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.954247] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 715.954446] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.954740] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28193bf0-ad6d-4539-aa59-053162b40fa8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.963630] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008b8e31-7a3d-4197-a575-ef6f135bede1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.984488] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7 could not be found. [ 715.984696] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 715.984875] env[62552]: INFO nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 715.985116] env[62552]: DEBUG oslo.service.loopingcall [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.985330] env[62552]: DEBUG nova.compute.manager [-] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 715.985423] env[62552]: DEBUG nova.network.neutron [-] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.005267] env[62552]: DEBUG nova.network.neutron [-] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.045141] env[62552]: DEBUG nova.scheduler.client.report [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 716.125289] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 716.149954] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.150227] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.150381] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.150735] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.150886] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.151043] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.151255] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.151411] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.151573] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.151765] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.152015] env[62552]: DEBUG nova.virt.hardware [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.152869] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f595ab-1e2d-4264-8028-a52d049361b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.163727] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b90dff0-fdf5-4f97-9e1b-d06e26457d40 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.237313] env[62552]: ERROR nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4cc71534-52d9-4238-8ab9-8384cf3ed303, please check neutron logs for more information. [ 716.237313] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.237313] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 716.237313] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.237313] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.237313] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.237313] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.237313] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.237313] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.237313] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 716.237313] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.237313] env[62552]: ERROR nova.compute.manager raise self.value [ 716.237313] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.237313] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.237313] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.237313] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.237775] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.237775] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.237775] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4cc71534-52d9-4238-8ab9-8384cf3ed303, please check neutron logs for more information. [ 716.237775] env[62552]: ERROR nova.compute.manager [ 716.237775] env[62552]: Traceback (most recent call last): [ 716.237775] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.237775] env[62552]: listener.cb(fileno) [ 716.237775] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.237775] env[62552]: result = function(*args, **kwargs) [ 716.237775] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.237775] env[62552]: return func(*args, **kwargs) [ 716.237775] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 716.237775] env[62552]: raise e [ 716.237775] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 716.237775] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 716.237775] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.237775] env[62552]: created_port_ids = self._update_ports_for_instance( [ 716.237775] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.237775] env[62552]: with excutils.save_and_reraise_exception(): [ 716.237775] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.237775] env[62552]: self.force_reraise() [ 716.237775] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.237775] env[62552]: raise self.value [ 716.237775] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.237775] env[62552]: updated_port = self._update_port( [ 716.237775] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.237775] env[62552]: _ensure_no_port_binding_failure(port) [ 716.237775] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.237775] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.238506] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 4cc71534-52d9-4238-8ab9-8384cf3ed303, please check neutron logs for more information. [ 716.238506] env[62552]: Removing descriptor: 15 [ 716.238506] env[62552]: ERROR nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4cc71534-52d9-4238-8ab9-8384cf3ed303, please check neutron logs for more information. [ 716.238506] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Traceback (most recent call last): [ 716.238506] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 716.238506] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] yield resources [ 716.238506] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 716.238506] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self.driver.spawn(context, instance, image_meta, [ 716.238506] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 716.238506] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.238506] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.238506] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] vm_ref = self.build_virtual_machine(instance, [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] for vif in network_info: [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] return self._sync_wrapper(fn, *args, **kwargs) [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self.wait() [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self[:] = self._gt.wait() [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] return self._exit_event.wait() [ 716.238819] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] result = hub.switch() [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] return self.greenlet.switch() [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] result = function(*args, **kwargs) [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] return func(*args, **kwargs) [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] raise e [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] nwinfo = self.network_api.allocate_for_instance( [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.239178] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] created_port_ids = self._update_ports_for_instance( [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] with excutils.save_and_reraise_exception(): [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self.force_reraise() [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] raise self.value [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] updated_port = self._update_port( [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] _ensure_no_port_binding_failure(port) [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.239502] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] raise exception.PortBindingFailed(port_id=port['id']) [ 716.239847] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] nova.exception.PortBindingFailed: Binding failed for port 4cc71534-52d9-4238-8ab9-8384cf3ed303, please check neutron logs for more information. [ 716.239847] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] [ 716.239847] env[62552]: INFO nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Terminating instance [ 716.508172] env[62552]: DEBUG nova.network.neutron [-] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.550212] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.550759] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 716.553799] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.514s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.555036] env[62552]: INFO nova.compute.claims [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.742938] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "refresh_cache-9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.743230] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquired lock "refresh_cache-9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.743342] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.011126] env[62552]: INFO nova.compute.manager [-] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Took 1.02 seconds to deallocate network for instance. [ 717.012609] env[62552]: DEBUG nova.compute.claims [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 717.012761] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.055716] env[62552]: DEBUG nova.compute.utils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.057019] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 717.057352] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 717.119212] env[62552]: DEBUG nova.policy [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02fad99a3a1f453aada7427baa061ba3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abb4ff37a7154d7da41e936b8358f93d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.270204] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.424594] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.521763] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Successfully created port: dab007d3-928e-46d5-8311-ae549b24faab {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.560570] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 717.841682] env[62552]: DEBUG nova.compute.manager [req-e18613d8-6072-4540-ae14-268881990d18 req-7576639b-353d-42e4-b9d1-3ef30b1e5ae7 service nova] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Received event network-changed-4cc71534-52d9-4238-8ab9-8384cf3ed303 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 717.841682] env[62552]: DEBUG nova.compute.manager [req-e18613d8-6072-4540-ae14-268881990d18 req-7576639b-353d-42e4-b9d1-3ef30b1e5ae7 service nova] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Refreshing instance network info cache due to event network-changed-4cc71534-52d9-4238-8ab9-8384cf3ed303. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 717.841682] env[62552]: DEBUG oslo_concurrency.lockutils [req-e18613d8-6072-4540-ae14-268881990d18 req-7576639b-353d-42e4-b9d1-3ef30b1e5ae7 service nova] Acquiring lock "refresh_cache-9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.930080] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Releasing lock "refresh_cache-9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.930500] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 717.930737] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 717.931247] env[62552]: DEBUG oslo_concurrency.lockutils [req-e18613d8-6072-4540-ae14-268881990d18 req-7576639b-353d-42e4-b9d1-3ef30b1e5ae7 service nova] Acquired lock "refresh_cache-9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.931427] env[62552]: DEBUG nova.network.neutron [req-e18613d8-6072-4540-ae14-268881990d18 req-7576639b-353d-42e4-b9d1-3ef30b1e5ae7 service nova] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Refreshing network info cache for port 4cc71534-52d9-4238-8ab9-8384cf3ed303 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 717.932731] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f99b6013-7200-461c-8d5d-4b6d47ef05b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.941568] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d812e4-4e6a-4771-bcbf-f5bb5b8ca9f3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.970828] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8 could not be found. [ 717.970828] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 717.970828] env[62552]: INFO nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 717.970828] env[62552]: DEBUG oslo.service.loopingcall [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.976981] env[62552]: DEBUG nova.compute.manager [-] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 717.977243] env[62552]: DEBUG nova.network.neutron [-] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.015867] env[62552]: DEBUG nova.network.neutron [-] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.031018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47385cfb-3269-49c6-a40f-a88929b2017f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.045748] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dfcad3f-26b0-4176-b50c-9bf3bf0360ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.117497] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88992bad-e6f8-420e-a0f0-b93810bfb022 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.126631] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecbf2e8a-046b-4f5a-b048-f7c956e38475 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.140881] env[62552]: DEBUG nova.compute.provider_tree [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.454398] env[62552]: DEBUG nova.network.neutron [req-e18613d8-6072-4540-ae14-268881990d18 req-7576639b-353d-42e4-b9d1-3ef30b1e5ae7 service nova] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.521920] env[62552]: DEBUG nova.network.neutron [-] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.615477] env[62552]: DEBUG nova.network.neutron [req-e18613d8-6072-4540-ae14-268881990d18 req-7576639b-353d-42e4-b9d1-3ef30b1e5ae7 service nova] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.619852] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 718.645374] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.645625] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.645789] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.645967] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.646124] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.646270] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.646468] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.646625] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.646790] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.646948] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.647131] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.647947] env[62552]: DEBUG nova.scheduler.client.report [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 718.651375] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e477090b-3304-4e25-82f3-efb3c52ede03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.660446] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87aeb83-76d8-4e76-aa55-a9b9db5a9df1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.887127] env[62552]: ERROR nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dab007d3-928e-46d5-8311-ae549b24faab, please check neutron logs for more information. [ 718.887127] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.887127] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.887127] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.887127] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.887127] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.887127] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.887127] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.887127] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.887127] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 718.887127] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.887127] env[62552]: ERROR nova.compute.manager raise self.value [ 718.887127] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.887127] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.887127] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.887127] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.887906] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.887906] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.887906] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dab007d3-928e-46d5-8311-ae549b24faab, please check neutron logs for more information. [ 718.887906] env[62552]: ERROR nova.compute.manager [ 718.887906] env[62552]: Traceback (most recent call last): [ 718.887906] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.887906] env[62552]: listener.cb(fileno) [ 718.887906] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.887906] env[62552]: result = function(*args, **kwargs) [ 718.887906] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.887906] env[62552]: return func(*args, **kwargs) [ 718.887906] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.887906] env[62552]: raise e [ 718.887906] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.887906] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 718.887906] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.887906] env[62552]: created_port_ids = self._update_ports_for_instance( [ 718.887906] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.887906] env[62552]: with excutils.save_and_reraise_exception(): [ 718.887906] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.887906] env[62552]: self.force_reraise() [ 718.887906] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.887906] env[62552]: raise self.value [ 718.887906] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.887906] env[62552]: updated_port = self._update_port( [ 718.887906] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.887906] env[62552]: _ensure_no_port_binding_failure(port) [ 718.887906] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.887906] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.888856] env[62552]: nova.exception.PortBindingFailed: Binding failed for port dab007d3-928e-46d5-8311-ae549b24faab, please check neutron logs for more information. [ 718.888856] env[62552]: Removing descriptor: 15 [ 718.888856] env[62552]: ERROR nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dab007d3-928e-46d5-8311-ae549b24faab, please check neutron logs for more information. [ 718.888856] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Traceback (most recent call last): [ 718.888856] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 718.888856] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] yield resources [ 718.888856] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 718.888856] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self.driver.spawn(context, instance, image_meta, [ 718.888856] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 718.888856] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.888856] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.888856] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] vm_ref = self.build_virtual_machine(instance, [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] for vif in network_info: [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] return self._sync_wrapper(fn, *args, **kwargs) [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self.wait() [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self[:] = self._gt.wait() [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] return self._exit_event.wait() [ 718.889260] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] result = hub.switch() [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] return self.greenlet.switch() [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] result = function(*args, **kwargs) [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] return func(*args, **kwargs) [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] raise e [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] nwinfo = self.network_api.allocate_for_instance( [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.889721] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] created_port_ids = self._update_ports_for_instance( [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] with excutils.save_and_reraise_exception(): [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self.force_reraise() [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] raise self.value [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] updated_port = self._update_port( [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] _ensure_no_port_binding_failure(port) [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.890169] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] raise exception.PortBindingFailed(port_id=port['id']) [ 718.890548] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] nova.exception.PortBindingFailed: Binding failed for port dab007d3-928e-46d5-8311-ae549b24faab, please check neutron logs for more information. [ 718.890548] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] [ 718.890548] env[62552]: INFO nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Terminating instance [ 719.024378] env[62552]: INFO nova.compute.manager [-] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Took 1.05 seconds to deallocate network for instance. [ 719.026585] env[62552]: DEBUG nova.compute.claims [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 719.026763] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.123280] env[62552]: DEBUG oslo_concurrency.lockutils [req-e18613d8-6072-4540-ae14-268881990d18 req-7576639b-353d-42e4-b9d1-3ef30b1e5ae7 service nova] Releasing lock "refresh_cache-9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.123383] env[62552]: DEBUG nova.compute.manager [req-e18613d8-6072-4540-ae14-268881990d18 req-7576639b-353d-42e4-b9d1-3ef30b1e5ae7 service nova] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Received event network-vif-deleted-4cc71534-52d9-4238-8ab9-8384cf3ed303 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 719.155406] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.155918] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 719.158647] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.174s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.160306] env[62552]: INFO nova.compute.claims [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.392447] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "refresh_cache-04289f4b-4342-4377-8f9c-3013eb646cc6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.392792] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "refresh_cache-04289f4b-4342-4377-8f9c-3013eb646cc6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.392867] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.664674] env[62552]: DEBUG nova.compute.utils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.669044] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 719.669044] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 719.715479] env[62552]: DEBUG nova.policy [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02fad99a3a1f453aada7427baa061ba3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abb4ff37a7154d7da41e936b8358f93d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.864692] env[62552]: DEBUG nova.compute.manager [req-f6dde520-1137-462f-ab4b-4e476e3a2fcc req-939bc3f8-6523-4cd9-a929-b50f7c49d19a service nova] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Received event network-changed-dab007d3-928e-46d5-8311-ae549b24faab {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 719.865605] env[62552]: DEBUG nova.compute.manager [req-f6dde520-1137-462f-ab4b-4e476e3a2fcc req-939bc3f8-6523-4cd9-a929-b50f7c49d19a service nova] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Refreshing instance network info cache due to event network-changed-dab007d3-928e-46d5-8311-ae549b24faab. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 719.865605] env[62552]: DEBUG oslo_concurrency.lockutils [req-f6dde520-1137-462f-ab4b-4e476e3a2fcc req-939bc3f8-6523-4cd9-a929-b50f7c49d19a service nova] Acquiring lock "refresh_cache-04289f4b-4342-4377-8f9c-3013eb646cc6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.915256] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.075474] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.109240] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Successfully created port: 857e2542-798b-477f-a0c8-9abb084fa66e {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 720.168463] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 720.528983] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59210932-5dc6-471c-8bab-093111cccec3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.536764] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9400fd-61b8-46cc-849f-2e408a2c3a60 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.571023] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8b3cdf-3f4a-488c-b4dc-87b8381a3d2e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.577944] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "refresh_cache-04289f4b-4342-4377-8f9c-3013eb646cc6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.578088] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 720.578275] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.579050] env[62552]: DEBUG oslo_concurrency.lockutils [req-f6dde520-1137-462f-ab4b-4e476e3a2fcc req-939bc3f8-6523-4cd9-a929-b50f7c49d19a service nova] Acquired lock "refresh_cache-04289f4b-4342-4377-8f9c-3013eb646cc6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.579050] env[62552]: DEBUG nova.network.neutron [req-f6dde520-1137-462f-ab4b-4e476e3a2fcc req-939bc3f8-6523-4cd9-a929-b50f7c49d19a service nova] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Refreshing network info cache for port dab007d3-928e-46d5-8311-ae549b24faab {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 720.579751] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d776c080-5453-485f-b431-32313dd75b42 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.584020] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732843c8-70ee-4d08-9487-56b901bb4e7a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.596542] env[62552]: DEBUG nova.compute.provider_tree [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.601784] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a07fd8-d70d-48a3-8cc8-d928c8e083b8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.624097] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 04289f4b-4342-4377-8f9c-3013eb646cc6 could not be found. [ 720.624393] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.624505] env[62552]: INFO nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 720.624738] env[62552]: DEBUG oslo.service.loopingcall [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.624950] env[62552]: DEBUG nova.compute.manager [-] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 720.625073] env[62552]: DEBUG nova.network.neutron [-] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.649402] env[62552]: DEBUG nova.network.neutron [-] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.100861] env[62552]: DEBUG nova.scheduler.client.report [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 721.123996] env[62552]: DEBUG nova.network.neutron [req-f6dde520-1137-462f-ab4b-4e476e3a2fcc req-939bc3f8-6523-4cd9-a929-b50f7c49d19a service nova] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.151201] env[62552]: DEBUG nova.network.neutron [-] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.179043] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 721.208340] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 721.208451] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 721.208550] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.208740] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 721.208916] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.209089] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 721.209395] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 721.209497] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 721.209630] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 721.209805] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 721.209973] env[62552]: DEBUG nova.virt.hardware [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.210904] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1886bd1-e2fb-47b2-a78a-e1c9068862a9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.219045] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a676261e-762b-484d-ae88-7b529dd90b0a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.418800] env[62552]: ERROR nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. [ 721.418800] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 721.418800] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 721.418800] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 721.418800] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.418800] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 721.418800] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.418800] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 721.418800] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.418800] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 721.418800] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.418800] env[62552]: ERROR nova.compute.manager raise self.value [ 721.418800] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.418800] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 721.418800] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.418800] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 721.419256] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.419256] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 721.419256] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. [ 721.419256] env[62552]: ERROR nova.compute.manager [ 721.419256] env[62552]: Traceback (most recent call last): [ 721.419256] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 721.419256] env[62552]: listener.cb(fileno) [ 721.419256] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.419256] env[62552]: result = function(*args, **kwargs) [ 721.419256] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.419256] env[62552]: return func(*args, **kwargs) [ 721.419256] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 721.419256] env[62552]: raise e [ 721.419256] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 721.419256] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 721.419256] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.419256] env[62552]: created_port_ids = self._update_ports_for_instance( [ 721.419256] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.419256] env[62552]: with excutils.save_and_reraise_exception(): [ 721.419256] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.419256] env[62552]: self.force_reraise() [ 721.419256] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.419256] env[62552]: raise self.value [ 721.419256] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.419256] env[62552]: updated_port = self._update_port( [ 721.419256] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.419256] env[62552]: _ensure_no_port_binding_failure(port) [ 721.419256] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.419256] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 721.420066] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. [ 721.420066] env[62552]: Removing descriptor: 15 [ 721.420066] env[62552]: ERROR nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. [ 721.420066] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Traceback (most recent call last): [ 721.420066] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 721.420066] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] yield resources [ 721.420066] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 721.420066] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self.driver.spawn(context, instance, image_meta, [ 721.420066] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 721.420066] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.420066] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.420066] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] vm_ref = self.build_virtual_machine(instance, [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] for vif in network_info: [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] return self._sync_wrapper(fn, *args, **kwargs) [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self.wait() [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self[:] = self._gt.wait() [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] return self._exit_event.wait() [ 721.420372] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] result = hub.switch() [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] return self.greenlet.switch() [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] result = function(*args, **kwargs) [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] return func(*args, **kwargs) [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] raise e [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] nwinfo = self.network_api.allocate_for_instance( [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.421050] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] created_port_ids = self._update_ports_for_instance( [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] with excutils.save_and_reraise_exception(): [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self.force_reraise() [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] raise self.value [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] updated_port = self._update_port( [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] _ensure_no_port_binding_failure(port) [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.421488] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] raise exception.PortBindingFailed(port_id=port['id']) [ 721.421940] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] nova.exception.PortBindingFailed: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. [ 721.421940] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] [ 721.421940] env[62552]: INFO nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Terminating instance [ 721.442585] env[62552]: DEBUG nova.network.neutron [req-f6dde520-1137-462f-ab4b-4e476e3a2fcc req-939bc3f8-6523-4cd9-a929-b50f7c49d19a service nova] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.606854] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.607419] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 721.610527] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.627s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.611439] env[62552]: INFO nova.compute.claims [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.653938] env[62552]: INFO nova.compute.manager [-] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Took 1.03 seconds to deallocate network for instance. [ 721.656117] env[62552]: DEBUG nova.compute.claims [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.656305] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.893488] env[62552]: DEBUG nova.compute.manager [req-9efd2703-ea64-4449-8ade-f08350a7a8bd req-17c375da-18ec-427d-8229-9eeef976ac0a service nova] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Received event network-changed-857e2542-798b-477f-a0c8-9abb084fa66e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 721.893638] env[62552]: DEBUG nova.compute.manager [req-9efd2703-ea64-4449-8ade-f08350a7a8bd req-17c375da-18ec-427d-8229-9eeef976ac0a service nova] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Refreshing instance network info cache due to event network-changed-857e2542-798b-477f-a0c8-9abb084fa66e. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 721.893884] env[62552]: DEBUG oslo_concurrency.lockutils [req-9efd2703-ea64-4449-8ade-f08350a7a8bd req-17c375da-18ec-427d-8229-9eeef976ac0a service nova] Acquiring lock "refresh_cache-f8f37b3c-5e74-4459-9177-aacfac707616" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.894007] env[62552]: DEBUG oslo_concurrency.lockutils [req-9efd2703-ea64-4449-8ade-f08350a7a8bd req-17c375da-18ec-427d-8229-9eeef976ac0a service nova] Acquired lock "refresh_cache-f8f37b3c-5e74-4459-9177-aacfac707616" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.894154] env[62552]: DEBUG nova.network.neutron [req-9efd2703-ea64-4449-8ade-f08350a7a8bd req-17c375da-18ec-427d-8229-9eeef976ac0a service nova] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Refreshing network info cache for port 857e2542-798b-477f-a0c8-9abb084fa66e {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 721.924194] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "refresh_cache-f8f37b3c-5e74-4459-9177-aacfac707616" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.944506] env[62552]: DEBUG oslo_concurrency.lockutils [req-f6dde520-1137-462f-ab4b-4e476e3a2fcc req-939bc3f8-6523-4cd9-a929-b50f7c49d19a service nova] Releasing lock "refresh_cache-04289f4b-4342-4377-8f9c-3013eb646cc6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.944763] env[62552]: DEBUG nova.compute.manager [req-f6dde520-1137-462f-ab4b-4e476e3a2fcc req-939bc3f8-6523-4cd9-a929-b50f7c49d19a service nova] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Received event network-vif-deleted-dab007d3-928e-46d5-8311-ae549b24faab {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 722.117789] env[62552]: DEBUG nova.compute.utils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.121246] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 722.121246] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 722.194910] env[62552]: DEBUG nova.policy [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '827f04ab2fc54ac48f0ec55942a4df0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef662f2c1fb74c22ad7464f4f51a8898', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.418909] env[62552]: DEBUG nova.network.neutron [req-9efd2703-ea64-4449-8ade-f08350a7a8bd req-17c375da-18ec-427d-8229-9eeef976ac0a service nova] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.580701] env[62552]: DEBUG nova.network.neutron [req-9efd2703-ea64-4449-8ade-f08350a7a8bd req-17c375da-18ec-427d-8229-9eeef976ac0a service nova] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.620692] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 722.683148] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Successfully created port: d0cc2aec-0deb-4b82-8540-f3b8a0c39e69 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.988110] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6380ed-1e8f-4bd7-b504-407d8241c167 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.994868] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc3ed05-5d28-441d-9802-e529afc63597 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.025961] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef9415c-9d9e-4b63-9f6d-9b430bf3fb7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.033658] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7053b8-d620-480d-93f2-a1ace3048807 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.046691] env[62552]: DEBUG nova.compute.provider_tree [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.082529] env[62552]: DEBUG oslo_concurrency.lockutils [req-9efd2703-ea64-4449-8ade-f08350a7a8bd req-17c375da-18ec-427d-8229-9eeef976ac0a service nova] Releasing lock "refresh_cache-f8f37b3c-5e74-4459-9177-aacfac707616" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.082844] env[62552]: DEBUG nova.compute.manager [req-9efd2703-ea64-4449-8ade-f08350a7a8bd req-17c375da-18ec-427d-8229-9eeef976ac0a service nova] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Received event network-vif-deleted-857e2542-798b-477f-a0c8-9abb084fa66e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 723.083165] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "refresh_cache-f8f37b3c-5e74-4459-9177-aacfac707616" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.083372] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.555022] env[62552]: DEBUG nova.scheduler.client.report [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 723.618293] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.634183] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 723.663051] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.663307] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.663464] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.663644] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.663901] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.663971] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.664227] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.664427] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.664627] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.664835] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.665052] env[62552]: DEBUG nova.virt.hardware [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.665926] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ae2e5c-0ec5-4ebf-849c-367a70c4de76 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.673970] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b75fd58-ab0f-477d-8967-0b0e027804f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.801299] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.921064] env[62552]: DEBUG nova.compute.manager [req-59e3c6f1-95cf-45de-bdd8-90d042def382 req-379670cf-b69b-441e-9efa-62e0a4f5ef3c service nova] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Received event network-changed-d0cc2aec-0deb-4b82-8540-f3b8a0c39e69 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 723.921064] env[62552]: DEBUG nova.compute.manager [req-59e3c6f1-95cf-45de-bdd8-90d042def382 req-379670cf-b69b-441e-9efa-62e0a4f5ef3c service nova] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Refreshing instance network info cache due to event network-changed-d0cc2aec-0deb-4b82-8540-f3b8a0c39e69. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 723.921064] env[62552]: DEBUG oslo_concurrency.lockutils [req-59e3c6f1-95cf-45de-bdd8-90d042def382 req-379670cf-b69b-441e-9efa-62e0a4f5ef3c service nova] Acquiring lock "refresh_cache-2031fc3f-2c79-4f6e-b7eb-29235814141d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.921064] env[62552]: DEBUG oslo_concurrency.lockutils [req-59e3c6f1-95cf-45de-bdd8-90d042def382 req-379670cf-b69b-441e-9efa-62e0a4f5ef3c service nova] Acquired lock "refresh_cache-2031fc3f-2c79-4f6e-b7eb-29235814141d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.921064] env[62552]: DEBUG nova.network.neutron [req-59e3c6f1-95cf-45de-bdd8-90d042def382 req-379670cf-b69b-441e-9efa-62e0a4f5ef3c service nova] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Refreshing network info cache for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 723.960338] env[62552]: ERROR nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69, please check neutron logs for more information. [ 723.960338] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.960338] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 723.960338] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.960338] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.960338] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.960338] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.960338] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.960338] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.960338] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 723.960338] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.960338] env[62552]: ERROR nova.compute.manager raise self.value [ 723.960338] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.960338] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.960338] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.960338] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.960798] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.960798] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.960798] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69, please check neutron logs for more information. [ 723.960798] env[62552]: ERROR nova.compute.manager [ 723.960798] env[62552]: Traceback (most recent call last): [ 723.960798] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.960798] env[62552]: listener.cb(fileno) [ 723.960798] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.960798] env[62552]: result = function(*args, **kwargs) [ 723.960798] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.960798] env[62552]: return func(*args, **kwargs) [ 723.960798] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 723.960798] env[62552]: raise e [ 723.960798] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 723.960798] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 723.960798] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.960798] env[62552]: created_port_ids = self._update_ports_for_instance( [ 723.960798] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.960798] env[62552]: with excutils.save_and_reraise_exception(): [ 723.960798] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.960798] env[62552]: self.force_reraise() [ 723.960798] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.960798] env[62552]: raise self.value [ 723.960798] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.960798] env[62552]: updated_port = self._update_port( [ 723.960798] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.960798] env[62552]: _ensure_no_port_binding_failure(port) [ 723.960798] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.960798] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.961572] env[62552]: nova.exception.PortBindingFailed: Binding failed for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69, please check neutron logs for more information. [ 723.961572] env[62552]: Removing descriptor: 16 [ 723.961572] env[62552]: ERROR nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69, please check neutron logs for more information. [ 723.961572] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Traceback (most recent call last): [ 723.961572] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 723.961572] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] yield resources [ 723.961572] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 723.961572] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self.driver.spawn(context, instance, image_meta, [ 723.961572] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 723.961572] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.961572] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.961572] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] vm_ref = self.build_virtual_machine(instance, [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] for vif in network_info: [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] return self._sync_wrapper(fn, *args, **kwargs) [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self.wait() [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self[:] = self._gt.wait() [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] return self._exit_event.wait() [ 723.961885] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] result = hub.switch() [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] return self.greenlet.switch() [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] result = function(*args, **kwargs) [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] return func(*args, **kwargs) [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] raise e [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] nwinfo = self.network_api.allocate_for_instance( [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.962217] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] created_port_ids = self._update_ports_for_instance( [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] with excutils.save_and_reraise_exception(): [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self.force_reraise() [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] raise self.value [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] updated_port = self._update_port( [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] _ensure_no_port_binding_failure(port) [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.962567] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] raise exception.PortBindingFailed(port_id=port['id']) [ 723.963121] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] nova.exception.PortBindingFailed: Binding failed for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69, please check neutron logs for more information. [ 723.963121] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] [ 723.963121] env[62552]: INFO nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Terminating instance [ 724.058828] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.059160] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 724.062418] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.678s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.303952] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "refresh_cache-f8f37b3c-5e74-4459-9177-aacfac707616" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.304436] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 724.304642] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 724.304987] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16367d33-ca2b-44d6-876c-4225740ed897 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.314501] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb7ec71-8dde-4584-bf32-b543fc4e1437 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.335620] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f8f37b3c-5e74-4459-9177-aacfac707616 could not be found. [ 724.336811] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 724.336811] env[62552]: INFO nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Took 0.03 seconds to destroy the instance on the hypervisor. [ 724.336811] env[62552]: DEBUG oslo.service.loopingcall [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.336811] env[62552]: DEBUG nova.compute.manager [-] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 724.336811] env[62552]: DEBUG nova.network.neutron [-] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 724.351241] env[62552]: DEBUG nova.network.neutron [-] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.438918] env[62552]: DEBUG nova.network.neutron [req-59e3c6f1-95cf-45de-bdd8-90d042def382 req-379670cf-b69b-441e-9efa-62e0a4f5ef3c service nova] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.466672] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Acquiring lock "refresh_cache-2031fc3f-2c79-4f6e-b7eb-29235814141d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.485853] env[62552]: DEBUG nova.network.neutron [req-59e3c6f1-95cf-45de-bdd8-90d042def382 req-379670cf-b69b-441e-9efa-62e0a4f5ef3c service nova] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.567614] env[62552]: DEBUG nova.compute.utils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.571235] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 724.571405] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 724.632861] env[62552]: DEBUG nova.policy [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c732a4a15f9b4b16b241c09f4ba373e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3daa5253a0cc4538903b7cabc2f98fa2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.854804] env[62552]: DEBUG nova.network.neutron [-] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.938268] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4918c027-762c-4776-b8d4-121463834999 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.945979] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c874ff5-0fab-4887-9311-01d2c1ccbaec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.976293] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2d9d8a-84c7-432b-ba67-be0f43c5b106 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.983417] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2ff29b-ac43-48c9-8a2f-a0495976e07c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.988547] env[62552]: DEBUG oslo_concurrency.lockutils [req-59e3c6f1-95cf-45de-bdd8-90d042def382 req-379670cf-b69b-441e-9efa-62e0a4f5ef3c service nova] Releasing lock "refresh_cache-2031fc3f-2c79-4f6e-b7eb-29235814141d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.988915] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Acquired lock "refresh_cache-2031fc3f-2c79-4f6e-b7eb-29235814141d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.989108] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.999415] env[62552]: DEBUG nova.compute.provider_tree [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.021911] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Successfully created port: dffe4ce2-04bd-4831-b366-f11dfa6650ca {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.074107] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 725.360587] env[62552]: INFO nova.compute.manager [-] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Took 1.02 seconds to deallocate network for instance. [ 725.363195] env[62552]: DEBUG nova.compute.claims [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 725.363370] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.502630] env[62552]: DEBUG nova.scheduler.client.report [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 725.509974] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.603390] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.986028] env[62552]: DEBUG nova.compute.manager [req-6f19ae6e-3759-47ff-bf06-b9b511b7a932 req-964530ad-ed51-40fd-8ad8-146c01b8e55c service nova] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Received event network-vif-deleted-d0cc2aec-0deb-4b82-8540-f3b8a0c39e69 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 726.007355] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.945s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.008029] env[62552]: ERROR nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8323852d-b171-4826-ba44-55a07704e2b1, please check neutron logs for more information. [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Traceback (most recent call last): [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self.driver.spawn(context, instance, image_meta, [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] vm_ref = self.build_virtual_machine(instance, [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.008029] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] for vif in network_info: [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] return self._sync_wrapper(fn, *args, **kwargs) [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self.wait() [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self[:] = self._gt.wait() [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] return self._exit_event.wait() [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] result = hub.switch() [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.008585] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] return self.greenlet.switch() [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] result = function(*args, **kwargs) [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] return func(*args, **kwargs) [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] raise e [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] nwinfo = self.network_api.allocate_for_instance( [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] created_port_ids = self._update_ports_for_instance( [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] with excutils.save_and_reraise_exception(): [ 726.009065] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] self.force_reraise() [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] raise self.value [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] updated_port = self._update_port( [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] _ensure_no_port_binding_failure(port) [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] raise exception.PortBindingFailed(port_id=port['id']) [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] nova.exception.PortBindingFailed: Binding failed for port 8323852d-b171-4826-ba44-55a07704e2b1, please check neutron logs for more information. [ 726.009389] env[62552]: ERROR nova.compute.manager [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] [ 726.009728] env[62552]: DEBUG nova.compute.utils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Binding failed for port 8323852d-b171-4826-ba44-55a07704e2b1, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 726.009923] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.450s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.011264] env[62552]: INFO nova.compute.claims [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.016909] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Build of instance dd4c1e35-0397-4ee0-932d-2dc7f0870f84 was re-scheduled: Binding failed for port 8323852d-b171-4826-ba44-55a07704e2b1, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 726.016909] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 726.016909] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "refresh_cache-dd4c1e35-0397-4ee0-932d-2dc7f0870f84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.016909] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquired lock "refresh_cache-dd4c1e35-0397-4ee0-932d-2dc7f0870f84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.017143] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.083931] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 726.106304] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Releasing lock "refresh_cache-2031fc3f-2c79-4f6e-b7eb-29235814141d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.106808] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 726.107031] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.109250] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.109403] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.109579] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.109795] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.109946] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.110101] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.110303] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.110461] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.110625] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.110793] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.110953] env[62552]: DEBUG nova.virt.hardware [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.111220] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b0ba46c-9ded-4891-ac46-a8bc8d2dc583 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.114110] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff245d9b-6ffe-4c7e-94de-e797c69d1659 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.123018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b89d70-dcb0-4600-887a-3597aa561c2c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.129130] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d7722d-480f-47a1-81d2-253922897bd9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.152459] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2031fc3f-2c79-4f6e-b7eb-29235814141d could not be found. [ 726.152691] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 726.152875] env[62552]: INFO nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 726.153124] env[62552]: DEBUG oslo.service.loopingcall [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.153370] env[62552]: DEBUG nova.compute.manager [-] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 726.153475] env[62552]: DEBUG nova.network.neutron [-] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.186660] env[62552]: DEBUG nova.network.neutron [-] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.303205] env[62552]: ERROR nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dffe4ce2-04bd-4831-b366-f11dfa6650ca, please check neutron logs for more information. [ 726.303205] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.303205] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.303205] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.303205] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.303205] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.303205] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.303205] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.303205] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.303205] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 726.303205] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.303205] env[62552]: ERROR nova.compute.manager raise self.value [ 726.303205] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.303205] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.303205] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.303205] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.303715] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.303715] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.303715] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dffe4ce2-04bd-4831-b366-f11dfa6650ca, please check neutron logs for more information. [ 726.303715] env[62552]: ERROR nova.compute.manager [ 726.303715] env[62552]: Traceback (most recent call last): [ 726.303715] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.303715] env[62552]: listener.cb(fileno) [ 726.303715] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.303715] env[62552]: result = function(*args, **kwargs) [ 726.303715] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.303715] env[62552]: return func(*args, **kwargs) [ 726.303715] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 726.303715] env[62552]: raise e [ 726.303715] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.303715] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 726.303715] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.303715] env[62552]: created_port_ids = self._update_ports_for_instance( [ 726.303715] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.303715] env[62552]: with excutils.save_and_reraise_exception(): [ 726.303715] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.303715] env[62552]: self.force_reraise() [ 726.303715] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.303715] env[62552]: raise self.value [ 726.303715] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.303715] env[62552]: updated_port = self._update_port( [ 726.303715] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.303715] env[62552]: _ensure_no_port_binding_failure(port) [ 726.303715] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.303715] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.304575] env[62552]: nova.exception.PortBindingFailed: Binding failed for port dffe4ce2-04bd-4831-b366-f11dfa6650ca, please check neutron logs for more information. [ 726.304575] env[62552]: Removing descriptor: 16 [ 726.304575] env[62552]: ERROR nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dffe4ce2-04bd-4831-b366-f11dfa6650ca, please check neutron logs for more information. [ 726.304575] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Traceback (most recent call last): [ 726.304575] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 726.304575] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] yield resources [ 726.304575] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 726.304575] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self.driver.spawn(context, instance, image_meta, [ 726.304575] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 726.304575] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.304575] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.304575] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] vm_ref = self.build_virtual_machine(instance, [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] for vif in network_info: [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] return self._sync_wrapper(fn, *args, **kwargs) [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self.wait() [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self[:] = self._gt.wait() [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] return self._exit_event.wait() [ 726.304992] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] result = hub.switch() [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] return self.greenlet.switch() [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] result = function(*args, **kwargs) [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] return func(*args, **kwargs) [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] raise e [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] nwinfo = self.network_api.allocate_for_instance( [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.305410] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] created_port_ids = self._update_ports_for_instance( [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] with excutils.save_and_reraise_exception(): [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self.force_reraise() [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] raise self.value [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] updated_port = self._update_port( [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] _ensure_no_port_binding_failure(port) [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.305798] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] raise exception.PortBindingFailed(port_id=port['id']) [ 726.306171] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] nova.exception.PortBindingFailed: Binding failed for port dffe4ce2-04bd-4831-b366-f11dfa6650ca, please check neutron logs for more information. [ 726.306171] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] [ 726.306171] env[62552]: INFO nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Terminating instance [ 726.534976] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.602448] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.688865] env[62552]: DEBUG nova.network.neutron [-] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.808649] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-cf0dcba7-a36e-435c-8c34-d352b49a723b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.808850] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-cf0dcba7-a36e-435c-8c34-d352b49a723b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.809020] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.105459] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Releasing lock "refresh_cache-dd4c1e35-0397-4ee0-932d-2dc7f0870f84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.105694] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 727.105874] env[62552]: DEBUG nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 727.106075] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.125584] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.192664] env[62552]: INFO nova.compute.manager [-] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Took 1.04 seconds to deallocate network for instance. [ 727.194996] env[62552]: DEBUG nova.compute.claims [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 727.195195] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.347305] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.381767] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130265b2-5325-44db-980a-d258f387ff8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.389683] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c5f55b-2511-4e65-8def-3753f562c220 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.420219] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46277be5-0680-4afa-be20-b61ed7ba21a0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.427406] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9357e525-937f-404c-a1a3-be03e3754d19 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.441185] env[62552]: DEBUG nova.compute.provider_tree [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.476950] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.628533] env[62552]: DEBUG nova.network.neutron [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.944434] env[62552]: DEBUG nova.scheduler.client.report [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 727.979848] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-cf0dcba7-a36e-435c-8c34-d352b49a723b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.980322] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 727.980493] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.980809] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e865b759-2131-4efb-864a-23d53e01c6ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.989627] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1691cd5d-a4af-4e4e-b3e7-2b391d0ad645 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.010994] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cf0dcba7-a36e-435c-8c34-d352b49a723b could not be found. [ 728.011228] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.011408] env[62552]: INFO nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 728.011643] env[62552]: DEBUG oslo.service.loopingcall [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.011857] env[62552]: DEBUG nova.compute.manager [-] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 728.011956] env[62552]: DEBUG nova.network.neutron [-] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.026514] env[62552]: DEBUG nova.network.neutron [-] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.031927] env[62552]: DEBUG nova.compute.manager [req-29c4f706-d8fb-45d5-9bc3-760c089155af req-18055de3-68c6-4583-baae-b76a3b07f0f6 service nova] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Received event network-changed-dffe4ce2-04bd-4831-b366-f11dfa6650ca {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 728.031927] env[62552]: DEBUG nova.compute.manager [req-29c4f706-d8fb-45d5-9bc3-760c089155af req-18055de3-68c6-4583-baae-b76a3b07f0f6 service nova] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Refreshing instance network info cache due to event network-changed-dffe4ce2-04bd-4831-b366-f11dfa6650ca. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 728.031927] env[62552]: DEBUG oslo_concurrency.lockutils [req-29c4f706-d8fb-45d5-9bc3-760c089155af req-18055de3-68c6-4583-baae-b76a3b07f0f6 service nova] Acquiring lock "refresh_cache-cf0dcba7-a36e-435c-8c34-d352b49a723b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.032251] env[62552]: DEBUG oslo_concurrency.lockutils [req-29c4f706-d8fb-45d5-9bc3-760c089155af req-18055de3-68c6-4583-baae-b76a3b07f0f6 service nova] Acquired lock "refresh_cache-cf0dcba7-a36e-435c-8c34-d352b49a723b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.032251] env[62552]: DEBUG nova.network.neutron [req-29c4f706-d8fb-45d5-9bc3-760c089155af req-18055de3-68c6-4583-baae-b76a3b07f0f6 service nova] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Refreshing network info cache for port dffe4ce2-04bd-4831-b366-f11dfa6650ca {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 728.130902] env[62552]: INFO nova.compute.manager [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: dd4c1e35-0397-4ee0-932d-2dc7f0870f84] Took 1.02 seconds to deallocate network for instance. [ 728.449744] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.449744] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 728.452810] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.917s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.454270] env[62552]: INFO nova.compute.claims [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.529022] env[62552]: DEBUG nova.network.neutron [-] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.558493] env[62552]: DEBUG nova.network.neutron [req-29c4f706-d8fb-45d5-9bc3-760c089155af req-18055de3-68c6-4583-baae-b76a3b07f0f6 service nova] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.632366] env[62552]: DEBUG nova.network.neutron [req-29c4f706-d8fb-45d5-9bc3-760c089155af req-18055de3-68c6-4583-baae-b76a3b07f0f6 service nova] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.954484] env[62552]: DEBUG nova.compute.utils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.955892] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 728.956077] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 729.017633] env[62552]: DEBUG nova.policy [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '990650b40f7c4b48bdd33c7efe616c04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '243cc9b351044a01831c49d29407dd95', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.031826] env[62552]: INFO nova.compute.manager [-] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Took 1.02 seconds to deallocate network for instance. [ 729.034243] env[62552]: DEBUG nova.compute.claims [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 729.034421] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.135487] env[62552]: DEBUG oslo_concurrency.lockutils [req-29c4f706-d8fb-45d5-9bc3-760c089155af req-18055de3-68c6-4583-baae-b76a3b07f0f6 service nova] Releasing lock "refresh_cache-cf0dcba7-a36e-435c-8c34-d352b49a723b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.135761] env[62552]: DEBUG nova.compute.manager [req-29c4f706-d8fb-45d5-9bc3-760c089155af req-18055de3-68c6-4583-baae-b76a3b07f0f6 service nova] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Received event network-vif-deleted-dffe4ce2-04bd-4831-b366-f11dfa6650ca {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 729.166203] env[62552]: INFO nova.scheduler.client.report [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Deleted allocations for instance dd4c1e35-0397-4ee0-932d-2dc7f0870f84 [ 729.428508] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Successfully created port: e178ee4b-2ae3-4b07-8ca1-f886a5858c53 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.463019] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 729.675943] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fa3a2995-de40-40d2-b08a-862ea55c9117 tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "dd4c1e35-0397-4ee0-932d-2dc7f0870f84" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.232s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.836752] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0e4b01-9831-42be-b02c-fdf1fb16529c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.844558] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99a9841-1e7d-4a20-ac2c-d6fa5b16f205 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.873753] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41cfdd6-7117-4abc-a630-6bc15fbb23ac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.880653] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608e2e20-e04c-4363-aa5e-8db29acc79a2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.893377] env[62552]: DEBUG nova.compute.provider_tree [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.180219] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 730.376418] env[62552]: DEBUG nova.compute.manager [req-3a3f0550-c291-45e4-bbeb-4bc8f81594cf req-f2d625f5-20d2-483f-a445-1e90fb30ca46 service nova] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Received event network-changed-e178ee4b-2ae3-4b07-8ca1-f886a5858c53 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 730.376523] env[62552]: DEBUG nova.compute.manager [req-3a3f0550-c291-45e4-bbeb-4bc8f81594cf req-f2d625f5-20d2-483f-a445-1e90fb30ca46 service nova] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Refreshing instance network info cache due to event network-changed-e178ee4b-2ae3-4b07-8ca1-f886a5858c53. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 730.376691] env[62552]: DEBUG oslo_concurrency.lockutils [req-3a3f0550-c291-45e4-bbeb-4bc8f81594cf req-f2d625f5-20d2-483f-a445-1e90fb30ca46 service nova] Acquiring lock "refresh_cache-a9a16b57-603a-4d43-b834-e6ea62675f3b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.376828] env[62552]: DEBUG oslo_concurrency.lockutils [req-3a3f0550-c291-45e4-bbeb-4bc8f81594cf req-f2d625f5-20d2-483f-a445-1e90fb30ca46 service nova] Acquired lock "refresh_cache-a9a16b57-603a-4d43-b834-e6ea62675f3b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.376983] env[62552]: DEBUG nova.network.neutron [req-3a3f0550-c291-45e4-bbeb-4bc8f81594cf req-f2d625f5-20d2-483f-a445-1e90fb30ca46 service nova] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Refreshing network info cache for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 730.399280] env[62552]: DEBUG nova.scheduler.client.report [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 730.471433] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 730.508125] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.508375] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.508527] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.508703] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.508847] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.508987] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.509267] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.509429] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.509592] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.509776] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.509957] env[62552]: DEBUG nova.virt.hardware [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.510869] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78530155-6575-4cbd-913e-34282c592e07 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.519508] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ba87ea-2d3f-470b-80e0-4e52b988f804 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.621776] env[62552]: ERROR nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53, please check neutron logs for more information. [ 730.621776] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 730.621776] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.621776] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 730.621776] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.621776] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 730.621776] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.621776] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 730.621776] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.621776] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 730.621776] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.621776] env[62552]: ERROR nova.compute.manager raise self.value [ 730.621776] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.621776] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 730.621776] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.621776] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 730.622574] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.622574] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 730.622574] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53, please check neutron logs for more information. [ 730.622574] env[62552]: ERROR nova.compute.manager [ 730.622574] env[62552]: Traceback (most recent call last): [ 730.622574] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 730.622574] env[62552]: listener.cb(fileno) [ 730.622574] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.622574] env[62552]: result = function(*args, **kwargs) [ 730.622574] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.622574] env[62552]: return func(*args, **kwargs) [ 730.622574] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.622574] env[62552]: raise e [ 730.622574] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.622574] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 730.622574] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.622574] env[62552]: created_port_ids = self._update_ports_for_instance( [ 730.622574] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.622574] env[62552]: with excutils.save_and_reraise_exception(): [ 730.622574] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.622574] env[62552]: self.force_reraise() [ 730.622574] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.622574] env[62552]: raise self.value [ 730.622574] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.622574] env[62552]: updated_port = self._update_port( [ 730.622574] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.622574] env[62552]: _ensure_no_port_binding_failure(port) [ 730.622574] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.622574] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 730.624825] env[62552]: nova.exception.PortBindingFailed: Binding failed for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53, please check neutron logs for more information. [ 730.624825] env[62552]: Removing descriptor: 16 [ 730.624825] env[62552]: ERROR nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53, please check neutron logs for more information. [ 730.624825] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Traceback (most recent call last): [ 730.624825] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 730.624825] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] yield resources [ 730.624825] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 730.624825] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self.driver.spawn(context, instance, image_meta, [ 730.624825] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 730.624825] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.624825] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.624825] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] vm_ref = self.build_virtual_machine(instance, [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] for vif in network_info: [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] return self._sync_wrapper(fn, *args, **kwargs) [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self.wait() [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self[:] = self._gt.wait() [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] return self._exit_event.wait() [ 730.625679] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] result = hub.switch() [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] return self.greenlet.switch() [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] result = function(*args, **kwargs) [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] return func(*args, **kwargs) [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] raise e [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] nwinfo = self.network_api.allocate_for_instance( [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.626557] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] created_port_ids = self._update_ports_for_instance( [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] with excutils.save_and_reraise_exception(): [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self.force_reraise() [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] raise self.value [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] updated_port = self._update_port( [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] _ensure_no_port_binding_failure(port) [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.627168] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] raise exception.PortBindingFailed(port_id=port['id']) [ 730.627782] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] nova.exception.PortBindingFailed: Binding failed for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53, please check neutron logs for more information. [ 730.627782] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] [ 730.627782] env[62552]: INFO nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Terminating instance [ 730.697640] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.893089] env[62552]: DEBUG nova.network.neutron [req-3a3f0550-c291-45e4-bbeb-4bc8f81594cf req-f2d625f5-20d2-483f-a445-1e90fb30ca46 service nova] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.905512] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.906134] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 730.909317] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.954s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.910786] env[62552]: INFO nova.compute.claims [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.971557] env[62552]: DEBUG nova.network.neutron [req-3a3f0550-c291-45e4-bbeb-4bc8f81594cf req-f2d625f5-20d2-483f-a445-1e90fb30ca46 service nova] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.126857] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Acquiring lock "refresh_cache-a9a16b57-603a-4d43-b834-e6ea62675f3b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.415168] env[62552]: DEBUG nova.compute.utils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.419436] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 731.419436] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 731.466324] env[62552]: DEBUG nova.policy [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6515846d33e44bca942767f9e1d650b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2f57f72de0c4adc9754458f4e7040b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 731.473531] env[62552]: DEBUG oslo_concurrency.lockutils [req-3a3f0550-c291-45e4-bbeb-4bc8f81594cf req-f2d625f5-20d2-483f-a445-1e90fb30ca46 service nova] Releasing lock "refresh_cache-a9a16b57-603a-4d43-b834-e6ea62675f3b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.473936] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Acquired lock "refresh_cache-a9a16b57-603a-4d43-b834-e6ea62675f3b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.474141] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.864703] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Successfully created port: 3801e787-9577-43e3-810a-e8fbeca93bd2 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.919422] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 731.993048] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.114271] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.245310] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a25b633-a0e4-4b51-a748-ae7158ef75f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.251216] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da2fa08-1b60-4b2a-9e03-a656057dbc8f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.280699] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5407b24-762e-4fcf-b7ba-da543f541254 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.288038] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64179c2e-21ec-4942-9160-cc63fb1171da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.300995] env[62552]: DEBUG nova.compute.provider_tree [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.480771] env[62552]: DEBUG nova.compute.manager [req-a3de6f0f-2abe-4488-9e38-df530399f3fe req-14d74ce0-63f7-41d7-a4c6-d53aa1eecf3f service nova] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Received event network-vif-deleted-e178ee4b-2ae3-4b07-8ca1-f886a5858c53 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 732.620022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Releasing lock "refresh_cache-a9a16b57-603a-4d43-b834-e6ea62675f3b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.620022] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 732.620022] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.620022] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7dc89701-a576-4fa1-ab4b-9bda9a65712c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.627651] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607a243f-afac-4416-9908-55f2ca52b9ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.650614] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a9a16b57-603a-4d43-b834-e6ea62675f3b could not be found. [ 732.651080] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.651380] env[62552]: INFO nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 732.651711] env[62552]: DEBUG oslo.service.loopingcall [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.652026] env[62552]: DEBUG nova.compute.manager [-] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 732.652239] env[62552]: DEBUG nova.network.neutron [-] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.679836] env[62552]: DEBUG nova.network.neutron [-] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.804334] env[62552]: DEBUG nova.scheduler.client.report [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 732.936226] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 732.964202] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.964446] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.965473] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.965473] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.965473] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.967587] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.967587] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.967587] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.967587] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.967587] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.967752] env[62552]: DEBUG nova.virt.hardware [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.969043] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45df8a2a-ac54-4592-82c8-915fa1e12abf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.977748] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b7a2ca-d478-476a-a4f8-7875e0d592ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.182835] env[62552]: DEBUG nova.network.neutron [-] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.309518] env[62552]: ERROR nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3801e787-9577-43e3-810a-e8fbeca93bd2, please check neutron logs for more information. [ 733.309518] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 733.309518] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 733.309518] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 733.309518] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.309518] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 733.309518] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.309518] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 733.309518] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.309518] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 733.309518] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.309518] env[62552]: ERROR nova.compute.manager raise self.value [ 733.309518] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.309518] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 733.309518] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.309518] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 733.309966] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.309966] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 733.309966] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3801e787-9577-43e3-810a-e8fbeca93bd2, please check neutron logs for more information. [ 733.309966] env[62552]: ERROR nova.compute.manager [ 733.309966] env[62552]: Traceback (most recent call last): [ 733.309966] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 733.309966] env[62552]: listener.cb(fileno) [ 733.309966] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.309966] env[62552]: result = function(*args, **kwargs) [ 733.309966] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.309966] env[62552]: return func(*args, **kwargs) [ 733.309966] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 733.309966] env[62552]: raise e [ 733.309966] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 733.309966] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 733.309966] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.309966] env[62552]: created_port_ids = self._update_ports_for_instance( [ 733.309966] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.309966] env[62552]: with excutils.save_and_reraise_exception(): [ 733.309966] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.309966] env[62552]: self.force_reraise() [ 733.309966] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.309966] env[62552]: raise self.value [ 733.309966] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.309966] env[62552]: updated_port = self._update_port( [ 733.309966] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.309966] env[62552]: _ensure_no_port_binding_failure(port) [ 733.309966] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.309966] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 733.310873] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 3801e787-9577-43e3-810a-e8fbeca93bd2, please check neutron logs for more information. [ 733.310873] env[62552]: Removing descriptor: 16 [ 733.310873] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.311224] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 733.318366] env[62552]: ERROR nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3801e787-9577-43e3-810a-e8fbeca93bd2, please check neutron logs for more information. [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Traceback (most recent call last): [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] yield resources [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self.driver.spawn(context, instance, image_meta, [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] vm_ref = self.build_virtual_machine(instance, [ 733.318366] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] for vif in network_info: [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] return self._sync_wrapper(fn, *args, **kwargs) [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self.wait() [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self[:] = self._gt.wait() [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] return self._exit_event.wait() [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 733.318782] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] result = hub.switch() [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] return self.greenlet.switch() [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] result = function(*args, **kwargs) [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] return func(*args, **kwargs) [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] raise e [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] nwinfo = self.network_api.allocate_for_instance( [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] created_port_ids = self._update_ports_for_instance( [ 733.319249] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] with excutils.save_and_reraise_exception(): [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self.force_reraise() [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] raise self.value [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] updated_port = self._update_port( [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] _ensure_no_port_binding_failure(port) [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] raise exception.PortBindingFailed(port_id=port['id']) [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] nova.exception.PortBindingFailed: Binding failed for port 3801e787-9577-43e3-810a-e8fbeca93bd2, please check neutron logs for more information. [ 733.319659] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] [ 733.320087] env[62552]: INFO nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Terminating instance [ 733.320087] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.304s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.685886] env[62552]: INFO nova.compute.manager [-] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Took 1.03 seconds to deallocate network for instance. [ 733.689025] env[62552]: DEBUG nova.compute.claims [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 733.689025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.819367] env[62552]: DEBUG nova.compute.utils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.820798] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 733.820974] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 733.825468] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-5ae752f6-9fd4-46fd-a7d2-558af22e2943" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.825676] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-5ae752f6-9fd4-46fd-a7d2-558af22e2943" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.825859] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 733.879953] env[62552]: DEBUG nova.policy [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c33af6bf649c43e8a5c98f458a4e4183', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '837700d8d99947f6b2fda68cfae3c977', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.157051] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41786bde-a6c1-47fe-97f8-a33ce3f4f1a2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.165345] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b021904d-a339-4bd9-bf91-c65a1c68ffec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.195139] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724ad24c-2492-4167-b50d-4af3db941cc4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.202227] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6cb2f5-27b1-4d28-b114-b92b8c7d55a4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.215567] env[62552]: DEBUG nova.compute.provider_tree [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.304158] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Successfully created port: 80315e0d-eebf-4daf-9329-ae5472a35a2b {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.323916] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 734.370051] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.508093] env[62552]: DEBUG nova.compute.manager [req-a61e0c33-1cf1-4d5e-ac10-ce57c6705329 req-adf9c6da-eba5-46cf-9804-eaad706254cd service nova] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Received event network-changed-3801e787-9577-43e3-810a-e8fbeca93bd2 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 734.508189] env[62552]: DEBUG nova.compute.manager [req-a61e0c33-1cf1-4d5e-ac10-ce57c6705329 req-adf9c6da-eba5-46cf-9804-eaad706254cd service nova] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Refreshing instance network info cache due to event network-changed-3801e787-9577-43e3-810a-e8fbeca93bd2. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 734.508369] env[62552]: DEBUG oslo_concurrency.lockutils [req-a61e0c33-1cf1-4d5e-ac10-ce57c6705329 req-adf9c6da-eba5-46cf-9804-eaad706254cd service nova] Acquiring lock "refresh_cache-5ae752f6-9fd4-46fd-a7d2-558af22e2943" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.520592] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.721061] env[62552]: DEBUG nova.scheduler.client.report [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 735.023456] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-5ae752f6-9fd4-46fd-a7d2-558af22e2943" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.023953] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 735.024199] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.025194] env[62552]: DEBUG oslo_concurrency.lockutils [req-a61e0c33-1cf1-4d5e-ac10-ce57c6705329 req-adf9c6da-eba5-46cf-9804-eaad706254cd service nova] Acquired lock "refresh_cache-5ae752f6-9fd4-46fd-a7d2-558af22e2943" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.025377] env[62552]: DEBUG nova.network.neutron [req-a61e0c33-1cf1-4d5e-ac10-ce57c6705329 req-adf9c6da-eba5-46cf-9804-eaad706254cd service nova] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Refreshing network info cache for port 3801e787-9577-43e3-810a-e8fbeca93bd2 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 735.027077] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f68b04a-9296-43ac-a695-91a24016502d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.037343] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29cbe30-bc15-4a9d-9a2c-b0c8397df143 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.059537] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ae752f6-9fd4-46fd-a7d2-558af22e2943 could not be found. [ 735.059769] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.059955] env[62552]: INFO nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Took 0.04 seconds to destroy the instance on the hypervisor. [ 735.060205] env[62552]: DEBUG oslo.service.loopingcall [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.060407] env[62552]: DEBUG nova.compute.manager [-] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 735.060499] env[62552]: DEBUG nova.network.neutron [-] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.093857] env[62552]: DEBUG nova.network.neutron [-] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.224540] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.225210] env[62552]: ERROR nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f, please check neutron logs for more information. [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Traceback (most recent call last): [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self.driver.spawn(context, instance, image_meta, [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] vm_ref = self.build_virtual_machine(instance, [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.225210] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] for vif in network_info: [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] return self._sync_wrapper(fn, *args, **kwargs) [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self.wait() [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self[:] = self._gt.wait() [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] return self._exit_event.wait() [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] result = hub.switch() [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 735.225536] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] return self.greenlet.switch() [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] result = function(*args, **kwargs) [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] return func(*args, **kwargs) [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] raise e [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] nwinfo = self.network_api.allocate_for_instance( [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] created_port_ids = self._update_ports_for_instance( [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] with excutils.save_and_reraise_exception(): [ 735.225870] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] self.force_reraise() [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] raise self.value [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] updated_port = self._update_port( [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] _ensure_no_port_binding_failure(port) [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] raise exception.PortBindingFailed(port_id=port['id']) [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] nova.exception.PortBindingFailed: Binding failed for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f, please check neutron logs for more information. [ 735.226214] env[62552]: ERROR nova.compute.manager [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] [ 735.226498] env[62552]: DEBUG nova.compute.utils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Binding failed for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 735.228039] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Build of instance 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7 was re-scheduled: Binding failed for port 7221d21a-c271-4bb8-b1e9-821a44f22b4f, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 735.228558] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 735.228820] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Acquiring lock "refresh_cache-4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.229014] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Acquired lock "refresh_cache-4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.229257] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.230355] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.204s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.339023] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 735.368877] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.369146] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.369299] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.369531] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.369624] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.369828] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.370064] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.370230] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.370393] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.370550] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.370736] env[62552]: DEBUG nova.virt.hardware [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.371600] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafcd080-1572-406e-9b39-d74a32da22b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.382880] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f32d89-6c74-43c9-91e8-8f614100a385 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.554489] env[62552]: DEBUG nova.network.neutron [req-a61e0c33-1cf1-4d5e-ac10-ce57c6705329 req-adf9c6da-eba5-46cf-9804-eaad706254cd service nova] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.596786] env[62552]: DEBUG nova.network.neutron [-] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.713614] env[62552]: ERROR nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 80315e0d-eebf-4daf-9329-ae5472a35a2b, please check neutron logs for more information. [ 735.713614] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.713614] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 735.713614] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.713614] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.713614] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.713614] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.713614] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.713614] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.713614] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 735.713614] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.713614] env[62552]: ERROR nova.compute.manager raise self.value [ 735.713614] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.713614] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.713614] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.713614] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.714133] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.714133] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.714133] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 80315e0d-eebf-4daf-9329-ae5472a35a2b, please check neutron logs for more information. [ 735.714133] env[62552]: ERROR nova.compute.manager [ 735.714133] env[62552]: Traceback (most recent call last): [ 735.714133] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.714133] env[62552]: listener.cb(fileno) [ 735.714133] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.714133] env[62552]: result = function(*args, **kwargs) [ 735.714133] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.714133] env[62552]: return func(*args, **kwargs) [ 735.714133] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 735.714133] env[62552]: raise e [ 735.714133] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 735.714133] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 735.714133] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.714133] env[62552]: created_port_ids = self._update_ports_for_instance( [ 735.714133] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.714133] env[62552]: with excutils.save_and_reraise_exception(): [ 735.714133] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.714133] env[62552]: self.force_reraise() [ 735.714133] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.714133] env[62552]: raise self.value [ 735.714133] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.714133] env[62552]: updated_port = self._update_port( [ 735.714133] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.714133] env[62552]: _ensure_no_port_binding_failure(port) [ 735.714133] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.714133] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.715037] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 80315e0d-eebf-4daf-9329-ae5472a35a2b, please check neutron logs for more information. [ 735.715037] env[62552]: Removing descriptor: 16 [ 735.715037] env[62552]: ERROR nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 80315e0d-eebf-4daf-9329-ae5472a35a2b, please check neutron logs for more information. [ 735.715037] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Traceback (most recent call last): [ 735.715037] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 735.715037] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] yield resources [ 735.715037] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 735.715037] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self.driver.spawn(context, instance, image_meta, [ 735.715037] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 735.715037] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.715037] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.715037] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] vm_ref = self.build_virtual_machine(instance, [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] for vif in network_info: [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] return self._sync_wrapper(fn, *args, **kwargs) [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self.wait() [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self[:] = self._gt.wait() [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] return self._exit_event.wait() [ 735.715417] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] result = hub.switch() [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] return self.greenlet.switch() [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] result = function(*args, **kwargs) [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] return func(*args, **kwargs) [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] raise e [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] nwinfo = self.network_api.allocate_for_instance( [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.715809] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] created_port_ids = self._update_ports_for_instance( [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] with excutils.save_and_reraise_exception(): [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self.force_reraise() [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] raise self.value [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] updated_port = self._update_port( [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] _ensure_no_port_binding_failure(port) [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.716217] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] raise exception.PortBindingFailed(port_id=port['id']) [ 735.716830] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] nova.exception.PortBindingFailed: Binding failed for port 80315e0d-eebf-4daf-9329-ae5472a35a2b, please check neutron logs for more information. [ 735.716830] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] [ 735.716830] env[62552]: INFO nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Terminating instance [ 735.719270] env[62552]: DEBUG nova.network.neutron [req-a61e0c33-1cf1-4d5e-ac10-ce57c6705329 req-adf9c6da-eba5-46cf-9804-eaad706254cd service nova] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.753977] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.832266] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.052145] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de247428-85e9-483a-b4bb-ef7db6312e19 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.059447] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd0fc56-0f0c-42a7-a39a-6c9fec03ede4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.087678] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d121830-126a-4b22-8dc6-fa3505813ef8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.094484] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d248c3-6247-4260-bde5-1b915d8c6de4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.108081] env[62552]: INFO nova.compute.manager [-] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Took 1.05 seconds to deallocate network for instance. [ 736.108838] env[62552]: DEBUG nova.compute.provider_tree [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.111589] env[62552]: DEBUG nova.compute.claims [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.111759] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.222020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Acquiring lock "refresh_cache-cfdee9ce-8c90-40cc-a09e-2386c719c02e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.222020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Acquired lock "refresh_cache-cfdee9ce-8c90-40cc-a09e-2386c719c02e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.222020] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.223243] env[62552]: DEBUG oslo_concurrency.lockutils [req-a61e0c33-1cf1-4d5e-ac10-ce57c6705329 req-adf9c6da-eba5-46cf-9804-eaad706254cd service nova] Releasing lock "refresh_cache-5ae752f6-9fd4-46fd-a7d2-558af22e2943" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.223470] env[62552]: DEBUG nova.compute.manager [req-a61e0c33-1cf1-4d5e-ac10-ce57c6705329 req-adf9c6da-eba5-46cf-9804-eaad706254cd service nova] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Received event network-vif-deleted-3801e787-9577-43e3-810a-e8fbeca93bd2 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 736.335026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Releasing lock "refresh_cache-4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.335026] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 736.335026] env[62552]: DEBUG nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 736.335208] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 736.349938] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.558653] env[62552]: DEBUG nova.compute.manager [req-bce306f5-0796-4c2e-9572-181c3a7368cf req-76d0a9a4-a380-4209-bcad-e0ce75d652f5 service nova] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Received event network-changed-80315e0d-eebf-4daf-9329-ae5472a35a2b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 736.558861] env[62552]: DEBUG nova.compute.manager [req-bce306f5-0796-4c2e-9572-181c3a7368cf req-76d0a9a4-a380-4209-bcad-e0ce75d652f5 service nova] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Refreshing instance network info cache due to event network-changed-80315e0d-eebf-4daf-9329-ae5472a35a2b. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 736.559055] env[62552]: DEBUG oslo_concurrency.lockutils [req-bce306f5-0796-4c2e-9572-181c3a7368cf req-76d0a9a4-a380-4209-bcad-e0ce75d652f5 service nova] Acquiring lock "refresh_cache-cfdee9ce-8c90-40cc-a09e-2386c719c02e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.612710] env[62552]: DEBUG nova.scheduler.client.report [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 736.745423] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.809272] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.853159] env[62552]: DEBUG nova.network.neutron [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.117294] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.887s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.117938] env[62552]: ERROR nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4cc71534-52d9-4238-8ab9-8384cf3ed303, please check neutron logs for more information. [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Traceback (most recent call last): [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self.driver.spawn(context, instance, image_meta, [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] vm_ref = self.build_virtual_machine(instance, [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.117938] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] for vif in network_info: [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] return self._sync_wrapper(fn, *args, **kwargs) [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self.wait() [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self[:] = self._gt.wait() [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] return self._exit_event.wait() [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] result = hub.switch() [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.118380] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] return self.greenlet.switch() [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] result = function(*args, **kwargs) [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] return func(*args, **kwargs) [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] raise e [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] nwinfo = self.network_api.allocate_for_instance( [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] created_port_ids = self._update_ports_for_instance( [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] with excutils.save_and_reraise_exception(): [ 737.118776] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] self.force_reraise() [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] raise self.value [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] updated_port = self._update_port( [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] _ensure_no_port_binding_failure(port) [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] raise exception.PortBindingFailed(port_id=port['id']) [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] nova.exception.PortBindingFailed: Binding failed for port 4cc71534-52d9-4238-8ab9-8384cf3ed303, please check neutron logs for more information. [ 737.119196] env[62552]: ERROR nova.compute.manager [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] [ 737.119532] env[62552]: DEBUG nova.compute.utils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Binding failed for port 4cc71534-52d9-4238-8ab9-8384cf3ed303, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.120219] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.464s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.122931] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Build of instance 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8 was re-scheduled: Binding failed for port 4cc71534-52d9-4238-8ab9-8384cf3ed303, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 737.123362] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 737.123579] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquiring lock "refresh_cache-9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.123724] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Acquired lock "refresh_cache-9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.123883] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.311709] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Releasing lock "refresh_cache-cfdee9ce-8c90-40cc-a09e-2386c719c02e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.312498] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 737.312498] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.312690] env[62552]: DEBUG oslo_concurrency.lockutils [req-bce306f5-0796-4c2e-9572-181c3a7368cf req-76d0a9a4-a380-4209-bcad-e0ce75d652f5 service nova] Acquired lock "refresh_cache-cfdee9ce-8c90-40cc-a09e-2386c719c02e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.312859] env[62552]: DEBUG nova.network.neutron [req-bce306f5-0796-4c2e-9572-181c3a7368cf req-76d0a9a4-a380-4209-bcad-e0ce75d652f5 service nova] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Refreshing network info cache for port 80315e0d-eebf-4daf-9329-ae5472a35a2b {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 737.313957] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed6b4cb3-2a95-40db-ae14-3329187bfb42 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.323529] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e4209d-4bdf-4b27-92cf-9c89d0a599f3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.345131] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cfdee9ce-8c90-40cc-a09e-2386c719c02e could not be found. [ 737.345341] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 737.345521] env[62552]: INFO nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 737.345759] env[62552]: DEBUG oslo.service.loopingcall [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.345972] env[62552]: DEBUG nova.compute.manager [-] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 737.346078] env[62552]: DEBUG nova.network.neutron [-] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.356292] env[62552]: INFO nova.compute.manager [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] [instance: 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7] Took 1.02 seconds to deallocate network for instance. [ 737.360038] env[62552]: DEBUG nova.network.neutron [-] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.641491] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.734896] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.833381] env[62552]: DEBUG nova.network.neutron [req-bce306f5-0796-4c2e-9572-181c3a7368cf req-76d0a9a4-a380-4209-bcad-e0ce75d652f5 service nova] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.861954] env[62552]: DEBUG nova.network.neutron [-] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.904096] env[62552]: DEBUG nova.network.neutron [req-bce306f5-0796-4c2e-9572-181c3a7368cf req-76d0a9a4-a380-4209-bcad-e0ce75d652f5 service nova] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.939958] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b27ed7-d8bf-4883-9a11-ded44a2b45be {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.947755] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cdd4cc3-acc5-45ff-a94c-0c26df1ab7b8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.978021] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4134ad5a-9f87-48bd-87f5-3a4afa332038 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.985498] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d999e5a-13d7-43aa-a8bc-f8f300b9baed {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.999594] env[62552]: DEBUG nova.compute.provider_tree [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.237263] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Releasing lock "refresh_cache-9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.237524] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 738.237707] env[62552]: DEBUG nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 738.237877] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.258205] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.366709] env[62552]: INFO nova.compute.manager [-] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Took 1.02 seconds to deallocate network for instance. [ 738.368941] env[62552]: DEBUG nova.compute.claims [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 738.369135] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.384528] env[62552]: INFO nova.scheduler.client.report [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Deleted allocations for instance 4c10ee2a-cc97-4ca9-bff6-776bfc476cc7 [ 738.406349] env[62552]: DEBUG oslo_concurrency.lockutils [req-bce306f5-0796-4c2e-9572-181c3a7368cf req-76d0a9a4-a380-4209-bcad-e0ce75d652f5 service nova] Releasing lock "refresh_cache-cfdee9ce-8c90-40cc-a09e-2386c719c02e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.406609] env[62552]: DEBUG nova.compute.manager [req-bce306f5-0796-4c2e-9572-181c3a7368cf req-76d0a9a4-a380-4209-bcad-e0ce75d652f5 service nova] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Received event network-vif-deleted-80315e0d-eebf-4daf-9329-ae5472a35a2b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 738.502700] env[62552]: DEBUG nova.scheduler.client.report [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 738.761596] env[62552]: DEBUG nova.network.neutron [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.892775] env[62552]: DEBUG oslo_concurrency.lockutils [None req-03fe5d6b-7b48-49e6-9287-d67785e03368 tempest-ServersNegativeTestJSON-207776419 tempest-ServersNegativeTestJSON-207776419-project-member] Lock "4c10ee2a-cc97-4ca9-bff6-776bfc476cc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.320s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.008487] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.888s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.009173] env[62552]: ERROR nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dab007d3-928e-46d5-8311-ae549b24faab, please check neutron logs for more information. [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Traceback (most recent call last): [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self.driver.spawn(context, instance, image_meta, [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] vm_ref = self.build_virtual_machine(instance, [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.009173] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] for vif in network_info: [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] return self._sync_wrapper(fn, *args, **kwargs) [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self.wait() [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self[:] = self._gt.wait() [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] return self._exit_event.wait() [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] result = hub.switch() [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.009527] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] return self.greenlet.switch() [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] result = function(*args, **kwargs) [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] return func(*args, **kwargs) [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] raise e [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] nwinfo = self.network_api.allocate_for_instance( [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] created_port_ids = self._update_ports_for_instance( [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] with excutils.save_and_reraise_exception(): [ 739.010065] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] self.force_reraise() [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] raise self.value [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] updated_port = self._update_port( [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] _ensure_no_port_binding_failure(port) [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] raise exception.PortBindingFailed(port_id=port['id']) [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] nova.exception.PortBindingFailed: Binding failed for port dab007d3-928e-46d5-8311-ae549b24faab, please check neutron logs for more information. [ 739.010460] env[62552]: ERROR nova.compute.manager [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] [ 739.011036] env[62552]: DEBUG nova.compute.utils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Binding failed for port dab007d3-928e-46d5-8311-ae549b24faab, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.011428] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.648s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.014067] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Build of instance 04289f4b-4342-4377-8f9c-3013eb646cc6 was re-scheduled: Binding failed for port dab007d3-928e-46d5-8311-ae549b24faab, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 739.014502] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 739.014725] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "refresh_cache-04289f4b-4342-4377-8f9c-3013eb646cc6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.014879] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "refresh_cache-04289f4b-4342-4377-8f9c-3013eb646cc6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.015038] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.265281] env[62552]: INFO nova.compute.manager [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] [instance: 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8] Took 1.03 seconds to deallocate network for instance. [ 739.396362] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 739.546789] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.662274] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.903131] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3d8776-ad75-4809-9c83-c287e8265bde {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.915610] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0537d6d1-4ffe-405d-a30f-7b329bda11ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.922079] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.946827] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00412575-d0c3-4581-aebf-91ff120642b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.954462] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86039df-7fe4-494f-9f2c-efe4a419e3e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.973380] env[62552]: DEBUG nova.compute.provider_tree [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.166300] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "refresh_cache-04289f4b-4342-4377-8f9c-3013eb646cc6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.166857] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 740.166857] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 740.167130] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.190946] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.298893] env[62552]: INFO nova.scheduler.client.report [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Deleted allocations for instance 9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8 [ 740.476868] env[62552]: DEBUG nova.scheduler.client.report [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 740.694036] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.809111] env[62552]: DEBUG oslo_concurrency.lockutils [None req-04923389-a0bb-4cc9-922d-dafef01d9fbd tempest-ListServerFiltersTestJSON-550164794 tempest-ListServerFiltersTestJSON-550164794-project-member] Lock "9377b49d-6d55-4c5f-8b64-3eea7bd8e5a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.613s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.982071] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.982071] env[62552]: ERROR nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. [ 740.982071] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Traceback (most recent call last): [ 740.982071] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 740.982071] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self.driver.spawn(context, instance, image_meta, [ 740.982071] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 740.982071] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.982071] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.982071] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] vm_ref = self.build_virtual_machine(instance, [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] for vif in network_info: [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] return self._sync_wrapper(fn, *args, **kwargs) [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self.wait() [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self[:] = self._gt.wait() [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] return self._exit_event.wait() [ 740.982547] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] result = hub.switch() [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] return self.greenlet.switch() [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] result = function(*args, **kwargs) [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] return func(*args, **kwargs) [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] raise e [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] nwinfo = self.network_api.allocate_for_instance( [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.982904] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] created_port_ids = self._update_ports_for_instance( [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] with excutils.save_and_reraise_exception(): [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] self.force_reraise() [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] raise self.value [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] updated_port = self._update_port( [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] _ensure_no_port_binding_failure(port) [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.983290] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] raise exception.PortBindingFailed(port_id=port['id']) [ 740.983614] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] nova.exception.PortBindingFailed: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. [ 740.983614] env[62552]: ERROR nova.compute.manager [instance: f8f37b3c-5e74-4459-9177-aacfac707616] [ 740.983614] env[62552]: DEBUG nova.compute.utils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.984169] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.789s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.987272] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Build of instance f8f37b3c-5e74-4459-9177-aacfac707616 was re-scheduled: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 740.987733] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 740.987967] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "refresh_cache-f8f37b3c-5e74-4459-9177-aacfac707616" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.988137] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "refresh_cache-f8f37b3c-5e74-4459-9177-aacfac707616" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.988299] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.199559] env[62552]: INFO nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 04289f4b-4342-4377-8f9c-3013eb646cc6] Took 1.03 seconds to deallocate network for instance. [ 741.311528] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 741.513861] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.667155] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.828535] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.849634] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ab6478-9473-43fb-be9d-215b4eb64b23 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.857014] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25b4139-a8b9-4647-a2ab-8820eb861a75 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.888554] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7611f27-3d70-4d9e-a1df-f28e06e38f7e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.896308] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8addd11-c00e-44cb-90f5-09e95a2adf54 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.909673] env[62552]: DEBUG nova.compute.provider_tree [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.175702] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "refresh_cache-f8f37b3c-5e74-4459-9177-aacfac707616" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.175956] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 742.176169] env[62552]: DEBUG nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 742.176342] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 742.192478] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.230252] env[62552]: INFO nova.scheduler.client.report [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Deleted allocations for instance 04289f4b-4342-4377-8f9c-3013eb646cc6 [ 742.413054] env[62552]: DEBUG nova.scheduler.client.report [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 742.695064] env[62552]: DEBUG nova.network.neutron [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.737654] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "04289f4b-4342-4377-8f9c-3013eb646cc6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.525s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.918125] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.918782] env[62552]: ERROR nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69, please check neutron logs for more information. [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Traceback (most recent call last): [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self.driver.spawn(context, instance, image_meta, [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] vm_ref = self.build_virtual_machine(instance, [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.918782] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] for vif in network_info: [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] return self._sync_wrapper(fn, *args, **kwargs) [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self.wait() [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self[:] = self._gt.wait() [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] return self._exit_event.wait() [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] result = hub.switch() [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.919179] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] return self.greenlet.switch() [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] result = function(*args, **kwargs) [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] return func(*args, **kwargs) [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] raise e [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] nwinfo = self.network_api.allocate_for_instance( [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] created_port_ids = self._update_ports_for_instance( [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] with excutils.save_and_reraise_exception(): [ 742.919586] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] self.force_reraise() [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] raise self.value [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] updated_port = self._update_port( [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] _ensure_no_port_binding_failure(port) [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] raise exception.PortBindingFailed(port_id=port['id']) [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] nova.exception.PortBindingFailed: Binding failed for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69, please check neutron logs for more information. [ 742.920077] env[62552]: ERROR nova.compute.manager [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] [ 742.920435] env[62552]: DEBUG nova.compute.utils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Binding failed for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.920770] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.886s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.923684] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Build of instance 2031fc3f-2c79-4f6e-b7eb-29235814141d was re-scheduled: Binding failed for port d0cc2aec-0deb-4b82-8540-f3b8a0c39e69, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 742.924124] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 742.924353] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Acquiring lock "refresh_cache-2031fc3f-2c79-4f6e-b7eb-29235814141d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.924497] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Acquired lock "refresh_cache-2031fc3f-2c79-4f6e-b7eb-29235814141d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.924654] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.197478] env[62552]: INFO nova.compute.manager [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: f8f37b3c-5e74-4459-9177-aacfac707616] Took 1.02 seconds to deallocate network for instance. [ 743.242535] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 743.456016] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.536665] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.710142] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9af0b7-19f6-4d7a-961e-e14d8ad04c00 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.717830] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ec7724-f519-4e15-ba7a-cd3815c0e78d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.748891] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecb40bf-f702-48ed-bcb0-9b39c403e874 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.761356] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f3835c-cede-440e-ab12-042210461a5c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.776229] env[62552]: DEBUG nova.compute.provider_tree [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.778170] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.038955] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Releasing lock "refresh_cache-2031fc3f-2c79-4f6e-b7eb-29235814141d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.039355] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 744.039517] env[62552]: DEBUG nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 744.039721] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.060679] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.280272] env[62552]: DEBUG nova.scheduler.client.report [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 744.490415] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "474dfa9e-d003-478f-b48f-09e0c0452ffd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.490589] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "474dfa9e-d003-478f-b48f-09e0c0452ffd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.516046] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.518648] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.563988] env[62552]: DEBUG nova.network.neutron [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.785663] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.865s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.786371] env[62552]: ERROR nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dffe4ce2-04bd-4831-b366-f11dfa6650ca, please check neutron logs for more information. [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Traceback (most recent call last): [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self.driver.spawn(context, instance, image_meta, [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] vm_ref = self.build_virtual_machine(instance, [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.786371] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] for vif in network_info: [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] return self._sync_wrapper(fn, *args, **kwargs) [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self.wait() [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self[:] = self._gt.wait() [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] return self._exit_event.wait() [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] result = hub.switch() [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 744.786758] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] return self.greenlet.switch() [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] result = function(*args, **kwargs) [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] return func(*args, **kwargs) [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] raise e [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] nwinfo = self.network_api.allocate_for_instance( [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] created_port_ids = self._update_ports_for_instance( [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] with excutils.save_and_reraise_exception(): [ 744.787169] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] self.force_reraise() [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] raise self.value [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] updated_port = self._update_port( [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] _ensure_no_port_binding_failure(port) [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] raise exception.PortBindingFailed(port_id=port['id']) [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] nova.exception.PortBindingFailed: Binding failed for port dffe4ce2-04bd-4831-b366-f11dfa6650ca, please check neutron logs for more information. [ 744.787538] env[62552]: ERROR nova.compute.manager [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] [ 744.787843] env[62552]: DEBUG nova.compute.utils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Binding failed for port dffe4ce2-04bd-4831-b366-f11dfa6650ca, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 744.788468] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.091s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.789994] env[62552]: INFO nova.compute.claims [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.792709] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Build of instance cf0dcba7-a36e-435c-8c34-d352b49a723b was re-scheduled: Binding failed for port dffe4ce2-04bd-4831-b366-f11dfa6650ca, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 744.793142] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 744.793366] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-cf0dcba7-a36e-435c-8c34-d352b49a723b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.793514] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-cf0dcba7-a36e-435c-8c34-d352b49a723b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.793669] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.066128] env[62552]: INFO nova.compute.manager [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] [instance: 2031fc3f-2c79-4f6e-b7eb-29235814141d] Took 1.03 seconds to deallocate network for instance. [ 745.234206] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ef544ad9-3849-4c62-b618-baa482d81f80 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "f8f37b3c-5e74-4459-9177-aacfac707616" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.970s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.234557] env[62552]: Traceback (most recent call last): [ 745.234606] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 745.234606] env[62552]: self.driver.spawn(context, instance, image_meta, [ 745.234606] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 745.234606] env[62552]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.234606] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.234606] env[62552]: vm_ref = self.build_virtual_machine(instance, [ 745.234606] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.234606] env[62552]: vif_infos = vmwarevif.get_vif_info(self._session, [ 745.234929] env[62552]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.234929] env[62552]: for vif in network_info: [ 745.234929] env[62552]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.234929] env[62552]: return self._sync_wrapper(fn, *args, **kwargs) [ 745.234929] env[62552]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.234929] env[62552]: self.wait() [ 745.234929] env[62552]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.234929] env[62552]: self[:] = self._gt.wait() [ 745.234929] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.234929] env[62552]: return self._exit_event.wait() [ 745.234929] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 745.234929] env[62552]: result = hub.switch() [ 745.234929] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 745.234929] env[62552]: return self.greenlet.switch() [ 745.234929] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.234929] env[62552]: result = function(*args, **kwargs) [ 745.234929] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 745.234929] env[62552]: return func(*args, **kwargs) [ 745.234929] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 745.234929] env[62552]: raise e [ 745.234929] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 745.234929] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 745.234929] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.234929] env[62552]: created_port_ids = self._update_ports_for_instance( [ 745.234929] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.234929] env[62552]: with excutils.save_and_reraise_exception(): [ 745.234929] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.234929] env[62552]: self.force_reraise() [ 745.234929] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.234929] env[62552]: raise self.value [ 745.234929] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.234929] env[62552]: updated_port = self._update_port( [ 745.234929] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.236023] env[62552]: _ensure_no_port_binding_failure(port) [ 745.236023] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.236023] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 745.236023] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. [ 745.236023] env[62552]: During handling of the above exception, another exception occurred: [ 745.236023] env[62552]: Traceback (most recent call last): [ 745.236023] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2462, in _do_build_and_run_instance [ 745.236023] env[62552]: self._build_and_run_instance(context, instance, image, [ 745.236023] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2754, in _build_and_run_instance [ 745.236023] env[62552]: raise exception.RescheduledException( [ 745.236023] env[62552]: nova.exception.RescheduledException: Build of instance f8f37b3c-5e74-4459-9177-aacfac707616 was re-scheduled: Binding failed for port 857e2542-798b-477f-a0c8-9abb084fa66e, please check neutron logs for more information. [ 745.236023] env[62552]: During handling of the above exception, another exception occurred: [ 745.236023] env[62552]: Traceback (most recent call last): [ 745.236023] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 745.236023] env[62552]: func(*args, **kwargs) [ 745.236023] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 745.236023] env[62552]: return func(*args, **kwargs) [ 745.236023] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 745.236023] env[62552]: return f(*args, **kwargs) [ 745.236023] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2353, in _locked_do_build_and_run_instance [ 745.236023] env[62552]: result = self._do_build_and_run_instance(*args, **kwargs) [ 745.236023] env[62552]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 745.236023] env[62552]: with excutils.save_and_reraise_exception(): [ 745.236023] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.236023] env[62552]: self.force_reraise() [ 745.236023] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.236023] env[62552]: raise self.value [ 745.236023] env[62552]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 745.236923] env[62552]: return f(self, context, *args, **kw) [ 745.236923] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 168, in decorated_function [ 745.236923] env[62552]: with excutils.save_and_reraise_exception(): [ 745.236923] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.236923] env[62552]: self.force_reraise() [ 745.236923] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.236923] env[62552]: raise self.value [ 745.236923] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 159, in decorated_function [ 745.236923] env[62552]: return function(self, context, *args, **kwargs) [ 745.236923] env[62552]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 745.236923] env[62552]: return function(self, context, *args, **kwargs) [ 745.236923] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 205, in decorated_function [ 745.236923] env[62552]: return function(self, context, *args, **kwargs) [ 745.236923] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2497, in _do_build_and_run_instance [ 745.236923] env[62552]: instance.save() [ 745.236923] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 745.236923] env[62552]: updates, result = self.indirection_api.object_action( [ 745.236923] env[62552]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 745.236923] env[62552]: return cctxt.call(context, 'object_action', objinst=objinst, [ 745.236923] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 745.236923] env[62552]: result = self.transport._send( [ 745.236923] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 745.236923] env[62552]: return self._driver.send(target, ctxt, message, [ 745.236923] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 745.236923] env[62552]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 745.236923] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 745.236923] env[62552]: raise result [ 745.236923] env[62552]: nova.exception_Remote.InstanceNotFound_Remote: Instance f8f37b3c-5e74-4459-9177-aacfac707616 could not be found. [ 745.236923] env[62552]: Traceback (most recent call last): [ 745.237936] env[62552]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 745.237936] env[62552]: return getattr(target, method)(*args, **kwargs) [ 745.237936] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 745.237936] env[62552]: return fn(self, *args, **kwargs) [ 745.237936] env[62552]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 745.237936] env[62552]: old_ref, inst_ref = db.instance_update_and_get_original( [ 745.237936] env[62552]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 745.237936] env[62552]: return f(*args, **kwargs) [ 745.237936] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 745.237936] env[62552]: with excutils.save_and_reraise_exception() as ectxt: [ 745.237936] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.237936] env[62552]: self.force_reraise() [ 745.237936] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.237936] env[62552]: raise self.value [ 745.237936] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 745.237936] env[62552]: return f(*args, **kwargs) [ 745.237936] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 745.237936] env[62552]: return f(context, *args, **kwargs) [ 745.237936] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 745.237936] env[62552]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 745.237936] env[62552]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 745.237936] env[62552]: raise exception.InstanceNotFound(instance_id=uuid) [ 745.237936] env[62552]: nova.exception.InstanceNotFound: Instance f8f37b3c-5e74-4459-9177-aacfac707616 could not be found. [ 745.312357] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.438803] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.737506] env[62552]: DEBUG nova.compute.manager [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 745.940651] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-cf0dcba7-a36e-435c-8c34-d352b49a723b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.940939] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 745.941162] env[62552]: DEBUG nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 745.941333] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.955873] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.077691] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506db8d4-4533-448b-900e-c6da22c89e01 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.085732] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c9e757-bd29-4d31-9068-0340990c87a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.091354] env[62552]: INFO nova.scheduler.client.report [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Deleted allocations for instance 2031fc3f-2c79-4f6e-b7eb-29235814141d [ 746.125062] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a53f7b4-ffce-4c14-b4b8-9b33ea1742ea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.132018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af237e5-14d3-48d5-8c90-01aa406f0824 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.146464] env[62552]: DEBUG nova.compute.provider_tree [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.256640] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.458170] env[62552]: DEBUG nova.network.neutron [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.624213] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c380055-495a-41e6-97e4-f3394a409cca tempest-TenantUsagesTestJSON-1127647039 tempest-TenantUsagesTestJSON-1127647039-project-member] Lock "2031fc3f-2c79-4f6e-b7eb-29235814141d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.350s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.649522] env[62552]: DEBUG nova.scheduler.client.report [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 746.960776] env[62552]: INFO nova.compute.manager [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: cf0dcba7-a36e-435c-8c34-d352b49a723b] Took 1.02 seconds to deallocate network for instance. [ 747.129410] env[62552]: DEBUG nova.compute.manager [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 747.157019] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.157019] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 747.158179] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.469s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.653074] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.664024] env[62552]: DEBUG nova.compute.utils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.669509] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 747.669755] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.743972] env[62552]: DEBUG nova.policy [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46ca7a5363804d35935dd58e98b317c5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c881b4de4394e1ca338f4b0a7864854', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.991839] env[62552]: INFO nova.scheduler.client.report [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted allocations for instance cf0dcba7-a36e-435c-8c34-d352b49a723b [ 748.008227] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14306597-da1b-4905-aac4-b1c7ce66eeec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.016544] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47583b7-e234-4f0a-b311-ba3b16a8dc33 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.048597] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3c6551-19c3-4676-844d-b7668336c504 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.056080] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f03b8d-f486-4738-96a4-a111c644e92d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.070861] env[62552]: DEBUG nova.compute.provider_tree [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.167118] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 748.296865] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Successfully created port: 365eabca-4338-4379-8328-135cccda66fa {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.512253] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c9b3932d-8af1-4387-b3fa-a8763075097a tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "cf0dcba7-a36e-435c-8c34-d352b49a723b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.354s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.576033] env[62552]: DEBUG nova.scheduler.client.report [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 749.016091] env[62552]: DEBUG nova.compute.manager [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 749.079513] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.921s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.079624] env[62552]: ERROR nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53, please check neutron logs for more information. [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Traceback (most recent call last): [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self.driver.spawn(context, instance, image_meta, [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] vm_ref = self.build_virtual_machine(instance, [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.079624] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] for vif in network_info: [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] return self._sync_wrapper(fn, *args, **kwargs) [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self.wait() [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self[:] = self._gt.wait() [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] return self._exit_event.wait() [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] result = hub.switch() [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.080050] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] return self.greenlet.switch() [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] result = function(*args, **kwargs) [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] return func(*args, **kwargs) [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] raise e [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] nwinfo = self.network_api.allocate_for_instance( [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] created_port_ids = self._update_ports_for_instance( [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] with excutils.save_and_reraise_exception(): [ 749.080410] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] self.force_reraise() [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] raise self.value [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] updated_port = self._update_port( [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] _ensure_no_port_binding_failure(port) [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] raise exception.PortBindingFailed(port_id=port['id']) [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] nova.exception.PortBindingFailed: Binding failed for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53, please check neutron logs for more information. [ 749.080763] env[62552]: ERROR nova.compute.manager [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] [ 749.081091] env[62552]: DEBUG nova.compute.utils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Binding failed for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.081536] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.970s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.087436] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Build of instance a9a16b57-603a-4d43-b834-e6ea62675f3b was re-scheduled: Binding failed for port e178ee4b-2ae3-4b07-8ca1-f886a5858c53, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 749.087870] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 749.089560] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Acquiring lock "refresh_cache-a9a16b57-603a-4d43-b834-e6ea62675f3b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.089842] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Acquired lock "refresh_cache-a9a16b57-603a-4d43-b834-e6ea62675f3b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.090507] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.177363] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 749.198116] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.198370] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.198528] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.198706] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.198850] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.198993] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.199347] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.199566] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.199794] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.200145] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.200280] env[62552]: DEBUG nova.virt.hardware [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.201226] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a4d98b-0ff8-4cd3-84e1-8fa8a8d4ee3f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.209178] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e8884f-e48d-443e-ba96-38d2773ff0aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.534637] env[62552]: DEBUG nova.compute.manager [req-ac4c4603-df9c-4098-a3e2-23a1c74e05b5 req-e6155346-410a-48de-bdd9-2e595ae600e9 service nova] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Received event network-changed-365eabca-4338-4379-8328-135cccda66fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 749.534844] env[62552]: DEBUG nova.compute.manager [req-ac4c4603-df9c-4098-a3e2-23a1c74e05b5 req-e6155346-410a-48de-bdd9-2e595ae600e9 service nova] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Refreshing instance network info cache due to event network-changed-365eabca-4338-4379-8328-135cccda66fa. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 749.535073] env[62552]: DEBUG oslo_concurrency.lockutils [req-ac4c4603-df9c-4098-a3e2-23a1c74e05b5 req-e6155346-410a-48de-bdd9-2e595ae600e9 service nova] Acquiring lock "refresh_cache-e8ec980c-0186-48cb-84b9-af902f23d2fb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.535221] env[62552]: DEBUG oslo_concurrency.lockutils [req-ac4c4603-df9c-4098-a3e2-23a1c74e05b5 req-e6155346-410a-48de-bdd9-2e595ae600e9 service nova] Acquired lock "refresh_cache-e8ec980c-0186-48cb-84b9-af902f23d2fb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.535378] env[62552]: DEBUG nova.network.neutron [req-ac4c4603-df9c-4098-a3e2-23a1c74e05b5 req-e6155346-410a-48de-bdd9-2e595ae600e9 service nova] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Refreshing network info cache for port 365eabca-4338-4379-8328-135cccda66fa {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.542073] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.623029] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.781020] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.797685] env[62552]: ERROR nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 365eabca-4338-4379-8328-135cccda66fa, please check neutron logs for more information. [ 749.797685] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.797685] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 749.797685] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.797685] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.797685] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.797685] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.797685] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.797685] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.797685] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 749.797685] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.797685] env[62552]: ERROR nova.compute.manager raise self.value [ 749.797685] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.797685] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.797685] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.797685] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.798213] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.798213] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.798213] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 365eabca-4338-4379-8328-135cccda66fa, please check neutron logs for more information. [ 749.798213] env[62552]: ERROR nova.compute.manager [ 749.798213] env[62552]: Traceback (most recent call last): [ 749.798213] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.798213] env[62552]: listener.cb(fileno) [ 749.798213] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.798213] env[62552]: result = function(*args, **kwargs) [ 749.798213] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.798213] env[62552]: return func(*args, **kwargs) [ 749.798213] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 749.798213] env[62552]: raise e [ 749.798213] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 749.798213] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 749.798213] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.798213] env[62552]: created_port_ids = self._update_ports_for_instance( [ 749.798213] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.798213] env[62552]: with excutils.save_and_reraise_exception(): [ 749.798213] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.798213] env[62552]: self.force_reraise() [ 749.798213] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.798213] env[62552]: raise self.value [ 749.798213] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.798213] env[62552]: updated_port = self._update_port( [ 749.798213] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.798213] env[62552]: _ensure_no_port_binding_failure(port) [ 749.798213] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.798213] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.799166] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 365eabca-4338-4379-8328-135cccda66fa, please check neutron logs for more information. [ 749.799166] env[62552]: Removing descriptor: 15 [ 749.799166] env[62552]: ERROR nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 365eabca-4338-4379-8328-135cccda66fa, please check neutron logs for more information. [ 749.799166] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Traceback (most recent call last): [ 749.799166] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 749.799166] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] yield resources [ 749.799166] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 749.799166] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self.driver.spawn(context, instance, image_meta, [ 749.799166] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 749.799166] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.799166] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.799166] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] vm_ref = self.build_virtual_machine(instance, [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] for vif in network_info: [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] return self._sync_wrapper(fn, *args, **kwargs) [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self.wait() [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self[:] = self._gt.wait() [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] return self._exit_event.wait() [ 749.799547] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] result = hub.switch() [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] return self.greenlet.switch() [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] result = function(*args, **kwargs) [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] return func(*args, **kwargs) [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] raise e [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] nwinfo = self.network_api.allocate_for_instance( [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.799956] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] created_port_ids = self._update_ports_for_instance( [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] with excutils.save_and_reraise_exception(): [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self.force_reraise() [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] raise self.value [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] updated_port = self._update_port( [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] _ensure_no_port_binding_failure(port) [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.800394] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] raise exception.PortBindingFailed(port_id=port['id']) [ 749.800779] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] nova.exception.PortBindingFailed: Binding failed for port 365eabca-4338-4379-8328-135cccda66fa, please check neutron logs for more information. [ 749.800779] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] [ 749.800779] env[62552]: INFO nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Terminating instance [ 749.985638] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1d5a5d-3ee6-4d50-a28c-5c43eb40afec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.994696] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba52f17-025c-4d67-aa5e-80b444e312fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.025266] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624fabd9-4bd0-4993-873e-02c66456c74b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.033157] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022cdfc5-5788-42a5-ab0a-ac9864d38bac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.047243] env[62552]: DEBUG nova.compute.provider_tree [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.066444] env[62552]: DEBUG nova.network.neutron [req-ac4c4603-df9c-4098-a3e2-23a1c74e05b5 req-e6155346-410a-48de-bdd9-2e595ae600e9 service nova] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.097068] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "22357d4e-9771-477c-9fc3-fe3d76f6e902" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.097410] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.164675] env[62552]: DEBUG nova.network.neutron [req-ac4c4603-df9c-4098-a3e2-23a1c74e05b5 req-e6155346-410a-48de-bdd9-2e595ae600e9 service nova] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.283665] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Releasing lock "refresh_cache-a9a16b57-603a-4d43-b834-e6ea62675f3b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.284443] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 750.284443] env[62552]: DEBUG nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 750.284443] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.300991] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.312389] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Acquiring lock "refresh_cache-e8ec980c-0186-48cb-84b9-af902f23d2fb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.551101] env[62552]: DEBUG nova.scheduler.client.report [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 750.667546] env[62552]: DEBUG oslo_concurrency.lockutils [req-ac4c4603-df9c-4098-a3e2-23a1c74e05b5 req-e6155346-410a-48de-bdd9-2e595ae600e9 service nova] Releasing lock "refresh_cache-e8ec980c-0186-48cb-84b9-af902f23d2fb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.667995] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Acquired lock "refresh_cache-e8ec980c-0186-48cb-84b9-af902f23d2fb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.668199] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.805641] env[62552]: DEBUG nova.network.neutron [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.058513] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.977s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.059159] env[62552]: ERROR nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3801e787-9577-43e3-810a-e8fbeca93bd2, please check neutron logs for more information. [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Traceback (most recent call last): [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self.driver.spawn(context, instance, image_meta, [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] vm_ref = self.build_virtual_machine(instance, [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.059159] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] for vif in network_info: [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] return self._sync_wrapper(fn, *args, **kwargs) [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self.wait() [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self[:] = self._gt.wait() [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] return self._exit_event.wait() [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] result = hub.switch() [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.059551] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] return self.greenlet.switch() [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] result = function(*args, **kwargs) [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] return func(*args, **kwargs) [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] raise e [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] nwinfo = self.network_api.allocate_for_instance( [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] created_port_ids = self._update_ports_for_instance( [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] with excutils.save_and_reraise_exception(): [ 751.060020] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] self.force_reraise() [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] raise self.value [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] updated_port = self._update_port( [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] _ensure_no_port_binding_failure(port) [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] raise exception.PortBindingFailed(port_id=port['id']) [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] nova.exception.PortBindingFailed: Binding failed for port 3801e787-9577-43e3-810a-e8fbeca93bd2, please check neutron logs for more information. [ 751.060500] env[62552]: ERROR nova.compute.manager [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] [ 751.060858] env[62552]: DEBUG nova.compute.utils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Binding failed for port 3801e787-9577-43e3-810a-e8fbeca93bd2, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.061138] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.692s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.064048] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Build of instance 5ae752f6-9fd4-46fd-a7d2-558af22e2943 was re-scheduled: Binding failed for port 3801e787-9577-43e3-810a-e8fbeca93bd2, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 751.064483] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 751.064705] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-5ae752f6-9fd4-46fd-a7d2-558af22e2943" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.064849] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-5ae752f6-9fd4-46fd-a7d2-558af22e2943" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.065015] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.185201] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.304743] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.310504] env[62552]: INFO nova.compute.manager [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] [instance: a9a16b57-603a-4d43-b834-e6ea62675f3b] Took 1.03 seconds to deallocate network for instance. [ 751.574693] env[62552]: DEBUG nova.compute.manager [req-dcc0f3db-f780-4f54-872d-c86f2152632c req-c1ada99f-8ecc-46c7-ba5d-200d3c98e40f service nova] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Received event network-vif-deleted-365eabca-4338-4379-8328-135cccda66fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 751.592321] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.719035] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.807732] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Releasing lock "refresh_cache-e8ec980c-0186-48cb-84b9-af902f23d2fb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.808207] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 751.808918] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.808918] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0092edb3-b41b-494f-8bc4-2859e8e6d512 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.823117] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fb1583-5761-43ff-a395-58291a50192f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.847060] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8ec980c-0186-48cb-84b9-af902f23d2fb could not be found. [ 751.847316] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.847505] env[62552]: INFO nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 751.847750] env[62552]: DEBUG oslo.service.loopingcall [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.850300] env[62552]: DEBUG nova.compute.manager [-] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 751.850395] env[62552]: DEBUG nova.network.neutron [-] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.872722] env[62552]: DEBUG nova.network.neutron [-] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.961505] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f77f397-970c-4f24-9056-ecdbc168d8bb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.969496] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ab86ed-82f6-4941-8480-a7410e7c559b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.999930] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e233c06a-3128-49b3-8af3-4272bde068b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.007166] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c15d751-5002-49a0-a126-30291e772516 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.020226] env[62552]: DEBUG nova.compute.provider_tree [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.221479] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-5ae752f6-9fd4-46fd-a7d2-558af22e2943" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.221779] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 752.221925] env[62552]: DEBUG nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 752.222110] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.237223] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.333761] env[62552]: INFO nova.scheduler.client.report [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Deleted allocations for instance a9a16b57-603a-4d43-b834-e6ea62675f3b [ 752.374535] env[62552]: DEBUG nova.network.neutron [-] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.523374] env[62552]: DEBUG nova.scheduler.client.report [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 752.741771] env[62552]: DEBUG nova.network.neutron [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.841709] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2dd95a67-8c6d-4e84-ad8f-7747e20210d7 tempest-FloatingIPsAssociationTestJSON-1896855148 tempest-FloatingIPsAssociationTestJSON-1896855148-project-member] Lock "a9a16b57-603a-4d43-b834-e6ea62675f3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.752s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.879026] env[62552]: INFO nova.compute.manager [-] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Took 1.03 seconds to deallocate network for instance. [ 752.880682] env[62552]: DEBUG nova.compute.claims [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 752.880867] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.028417] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.029066] env[62552]: ERROR nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 80315e0d-eebf-4daf-9329-ae5472a35a2b, please check neutron logs for more information. [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Traceback (most recent call last): [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self.driver.spawn(context, instance, image_meta, [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] vm_ref = self.build_virtual_machine(instance, [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.029066] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] for vif in network_info: [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] return self._sync_wrapper(fn, *args, **kwargs) [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self.wait() [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self[:] = self._gt.wait() [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] return self._exit_event.wait() [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] result = hub.switch() [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 753.029466] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] return self.greenlet.switch() [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] result = function(*args, **kwargs) [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] return func(*args, **kwargs) [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] raise e [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] nwinfo = self.network_api.allocate_for_instance( [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] created_port_ids = self._update_ports_for_instance( [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] with excutils.save_and_reraise_exception(): [ 753.029870] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] self.force_reraise() [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] raise self.value [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] updated_port = self._update_port( [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] _ensure_no_port_binding_failure(port) [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] raise exception.PortBindingFailed(port_id=port['id']) [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] nova.exception.PortBindingFailed: Binding failed for port 80315e0d-eebf-4daf-9329-ae5472a35a2b, please check neutron logs for more information. [ 753.031613] env[62552]: ERROR nova.compute.manager [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] [ 753.031975] env[62552]: DEBUG nova.compute.utils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Binding failed for port 80315e0d-eebf-4daf-9329-ae5472a35a2b, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 753.033166] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.109s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.038285] env[62552]: INFO nova.compute.claims [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.038434] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Build of instance cfdee9ce-8c90-40cc-a09e-2386c719c02e was re-scheduled: Binding failed for port 80315e0d-eebf-4daf-9329-ae5472a35a2b, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 753.038911] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 753.039153] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Acquiring lock "refresh_cache-cfdee9ce-8c90-40cc-a09e-2386c719c02e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.039298] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Acquired lock "refresh_cache-cfdee9ce-8c90-40cc-a09e-2386c719c02e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.039614] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.244648] env[62552]: INFO nova.compute.manager [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 5ae752f6-9fd4-46fd-a7d2-558af22e2943] Took 1.02 seconds to deallocate network for instance. [ 753.345972] env[62552]: DEBUG nova.compute.manager [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 753.579371] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.706723] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.872288] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.212594] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Releasing lock "refresh_cache-cfdee9ce-8c90-40cc-a09e-2386c719c02e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.212594] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 754.212594] env[62552]: DEBUG nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 754.212594] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.249219] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.280593] env[62552]: INFO nova.scheduler.client.report [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleted allocations for instance 5ae752f6-9fd4-46fd-a7d2-558af22e2943 [ 754.410527] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e6216f-396b-4548-98f0-f730792e0f7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.418223] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d9c3f0-a8bf-4e16-b860-4e31c28079ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.456960] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41418ec8-14d2-4009-8a9d-d0559060c5fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.464184] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ef4848-23ab-45d0-a887-eed59b28b9ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.478362] env[62552]: DEBUG nova.compute.provider_tree [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.747142] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.747142] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.753137] env[62552]: DEBUG nova.network.neutron [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.792963] env[62552]: DEBUG oslo_concurrency.lockutils [None req-05bca6f9-0b31-45a1-9af0-41b45743485d tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "5ae752f6-9fd4-46fd-a7d2-558af22e2943" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.316s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.981545] env[62552]: DEBUG nova.scheduler.client.report [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 755.178217] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "4083ee43-ecea-4ea5-8923-42b348893824" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.178445] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "4083ee43-ecea-4ea5-8923-42b348893824" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.256697] env[62552]: INFO nova.compute.manager [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] [instance: cfdee9ce-8c90-40cc-a09e-2386c719c02e] Took 1.05 seconds to deallocate network for instance. [ 755.295720] env[62552]: DEBUG nova.compute.manager [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 755.487434] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.488034] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 755.492407] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.662s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.492407] env[62552]: INFO nova.compute.claims [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.824186] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.003010] env[62552]: DEBUG nova.compute.utils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.007877] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 756.008136] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 756.092759] env[62552]: DEBUG nova.policy [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '115e50fa02f3406488ec073cf627b5f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e31ae73234564af9bf84014f952e1018', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.292079] env[62552]: INFO nova.scheduler.client.report [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Deleted allocations for instance cfdee9ce-8c90-40cc-a09e-2386c719c02e [ 756.509178] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 756.688525] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Successfully created port: 4ef84cd8-05a5-4f97-8e3e-020d99a73896 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.802057] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5dc89eaa-b0de-4392-bcfb-7e2d59f14b31 tempest-ServersV294TestFqdnHostnames-672963599 tempest-ServersV294TestFqdnHostnames-672963599-project-member] Lock "cfdee9ce-8c90-40cc-a09e-2386c719c02e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.847s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.891808] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abaec5be-d394-443f-84e6-8f8d15687ccd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.903011] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6e69bd-f378-4500-a1fb-ad7488601c48 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.933566] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7504433-28cc-49ec-9b53-37c6c0942bd0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.944997] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81a737c-483d-4d38-ad1d-d820b8366c30 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.959432] env[62552]: DEBUG nova.compute.provider_tree [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.305790] env[62552]: DEBUG nova.compute.manager [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 757.462408] env[62552]: DEBUG nova.scheduler.client.report [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 757.524313] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 757.548478] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.548726] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.548878] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.549068] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.549224] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.549528] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.549759] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.549952] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.550140] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.550304] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.550474] env[62552]: DEBUG nova.virt.hardware [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.551689] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4852e35-88ad-4ce9-90fd-24f7cd4e23c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.560512] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30bff3a-ebd2-4faf-a230-5f42275f7f89 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.836680] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.968469] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.968526] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 757.971280] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.193s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.973048] env[62552]: INFO nova.compute.claims [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 758.478769] env[62552]: DEBUG nova.compute.utils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.479445] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 758.479627] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 758.577461] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.577642] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.610671] env[62552]: DEBUG nova.policy [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '812d8740af2046c98a0d388ac56d30f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9cc03743e2b47a2bca8e93127d22fa4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.985474] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 759.030387] env[62552]: DEBUG nova.compute.manager [req-2629b7d0-42b2-4790-be4b-a250731d6310 req-0b0af3e7-0ba8-4723-aa79-0fe5b8d1518b service nova] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Received event network-changed-4ef84cd8-05a5-4f97-8e3e-020d99a73896 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 759.030555] env[62552]: DEBUG nova.compute.manager [req-2629b7d0-42b2-4790-be4b-a250731d6310 req-0b0af3e7-0ba8-4723-aa79-0fe5b8d1518b service nova] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Refreshing instance network info cache due to event network-changed-4ef84cd8-05a5-4f97-8e3e-020d99a73896. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 759.030773] env[62552]: DEBUG oslo_concurrency.lockutils [req-2629b7d0-42b2-4790-be4b-a250731d6310 req-0b0af3e7-0ba8-4723-aa79-0fe5b8d1518b service nova] Acquiring lock "refresh_cache-ab363b52-ade2-45c4-ba3a-ae46ffaf831e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.030936] env[62552]: DEBUG oslo_concurrency.lockutils [req-2629b7d0-42b2-4790-be4b-a250731d6310 req-0b0af3e7-0ba8-4723-aa79-0fe5b8d1518b service nova] Acquired lock "refresh_cache-ab363b52-ade2-45c4-ba3a-ae46ffaf831e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.031142] env[62552]: DEBUG nova.network.neutron [req-2629b7d0-42b2-4790-be4b-a250731d6310 req-0b0af3e7-0ba8-4723-aa79-0fe5b8d1518b service nova] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Refreshing network info cache for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 759.392935] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486bc89a-19f0-4653-9112-a9425aa4257f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.403404] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401c0a4b-4af3-49fe-a667-1cbdb8c28b65 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.442778] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c500a8a-a7d1-4472-b966-2e8df3222aec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.452073] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5fc54a-a89a-4077-9581-9ed72238a15a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.466714] env[62552]: DEBUG nova.compute.provider_tree [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.473542] env[62552]: ERROR nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896, please check neutron logs for more information. [ 759.473542] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 759.473542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 759.473542] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 759.473542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.473542] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 759.473542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.473542] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 759.473542] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.473542] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 759.473542] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.473542] env[62552]: ERROR nova.compute.manager raise self.value [ 759.473542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.473542] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 759.473542] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.473542] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 759.474117] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.474117] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 759.474117] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896, please check neutron logs for more information. [ 759.474117] env[62552]: ERROR nova.compute.manager [ 759.474117] env[62552]: Traceback (most recent call last): [ 759.474117] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 759.474117] env[62552]: listener.cb(fileno) [ 759.474117] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.474117] env[62552]: result = function(*args, **kwargs) [ 759.474117] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.474117] env[62552]: return func(*args, **kwargs) [ 759.474117] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 759.474117] env[62552]: raise e [ 759.474117] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 759.474117] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 759.474117] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.474117] env[62552]: created_port_ids = self._update_ports_for_instance( [ 759.474117] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.474117] env[62552]: with excutils.save_and_reraise_exception(): [ 759.474117] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.474117] env[62552]: self.force_reraise() [ 759.474117] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.474117] env[62552]: raise self.value [ 759.474117] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.474117] env[62552]: updated_port = self._update_port( [ 759.474117] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.474117] env[62552]: _ensure_no_port_binding_failure(port) [ 759.474117] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.474117] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 759.475351] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896, please check neutron logs for more information. [ 759.475351] env[62552]: Removing descriptor: 15 [ 759.475351] env[62552]: ERROR nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896, please check neutron logs for more information. [ 759.475351] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Traceback (most recent call last): [ 759.475351] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 759.475351] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] yield resources [ 759.475351] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 759.475351] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self.driver.spawn(context, instance, image_meta, [ 759.475351] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 759.475351] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.475351] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.475351] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] vm_ref = self.build_virtual_machine(instance, [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] for vif in network_info: [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] return self._sync_wrapper(fn, *args, **kwargs) [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self.wait() [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self[:] = self._gt.wait() [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] return self._exit_event.wait() [ 759.475771] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] result = hub.switch() [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] return self.greenlet.switch() [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] result = function(*args, **kwargs) [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] return func(*args, **kwargs) [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] raise e [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] nwinfo = self.network_api.allocate_for_instance( [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.476197] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] created_port_ids = self._update_ports_for_instance( [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] with excutils.save_and_reraise_exception(): [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self.force_reraise() [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] raise self.value [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] updated_port = self._update_port( [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] _ensure_no_port_binding_failure(port) [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.476619] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] raise exception.PortBindingFailed(port_id=port['id']) [ 759.478156] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] nova.exception.PortBindingFailed: Binding failed for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896, please check neutron logs for more information. [ 759.478156] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] [ 759.478156] env[62552]: INFO nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Terminating instance [ 759.478156] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Successfully created port: 5e01d6b7-47ac-44d6-bf2f-097cb2207614 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.556381] env[62552]: DEBUG nova.network.neutron [req-2629b7d0-42b2-4790-be4b-a250731d6310 req-0b0af3e7-0ba8-4723-aa79-0fe5b8d1518b service nova] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.731581] env[62552]: DEBUG nova.network.neutron [req-2629b7d0-42b2-4790-be4b-a250731d6310 req-0b0af3e7-0ba8-4723-aa79-0fe5b8d1518b service nova] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.973461] env[62552]: DEBUG nova.scheduler.client.report [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 759.981501] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Acquiring lock "refresh_cache-ab363b52-ade2-45c4-ba3a-ae46ffaf831e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.998221] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 760.024220] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 760.024529] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 760.024699] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.024882] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 760.025039] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.025194] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 760.025400] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 760.025562] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 760.025729] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 760.025889] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 760.026071] env[62552]: DEBUG nova.virt.hardware [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.027424] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015c4bae-d8eb-4c04-bba6-a9004e9cdab6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.038017] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a7dd43-7ef9-4c76-bda2-50fba1816531 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.235314] env[62552]: DEBUG oslo_concurrency.lockutils [req-2629b7d0-42b2-4790-be4b-a250731d6310 req-0b0af3e7-0ba8-4723-aa79-0fe5b8d1518b service nova] Releasing lock "refresh_cache-ab363b52-ade2-45c4-ba3a-ae46ffaf831e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.235314] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Acquired lock "refresh_cache-ab363b52-ade2-45c4-ba3a-ae46ffaf831e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.235314] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.478406] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.478938] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 760.486038] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.226s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.486038] env[62552]: INFO nova.compute.claims [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.766289] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.937090] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.960727] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquiring lock "1cfeedac-f71e-42e4-a04f-8a0462c85907" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.960975] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "1cfeedac-f71e-42e4-a04f-8a0462c85907" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.988851] env[62552]: DEBUG nova.compute.utils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.990185] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 760.990347] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 761.060213] env[62552]: DEBUG nova.compute.manager [req-7e537c6e-fee1-4656-949d-f6fb52952a79 req-525f6c1f-6837-4e64-afdd-36fb58cbf4d5 service nova] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Received event network-vif-deleted-4ef84cd8-05a5-4f97-8e3e-020d99a73896 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 761.060418] env[62552]: DEBUG nova.compute.manager [req-7e537c6e-fee1-4656-949d-f6fb52952a79 req-525f6c1f-6837-4e64-afdd-36fb58cbf4d5 service nova] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Received event network-changed-5e01d6b7-47ac-44d6-bf2f-097cb2207614 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 761.060579] env[62552]: DEBUG nova.compute.manager [req-7e537c6e-fee1-4656-949d-f6fb52952a79 req-525f6c1f-6837-4e64-afdd-36fb58cbf4d5 service nova] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Refreshing instance network info cache due to event network-changed-5e01d6b7-47ac-44d6-bf2f-097cb2207614. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 761.060785] env[62552]: DEBUG oslo_concurrency.lockutils [req-7e537c6e-fee1-4656-949d-f6fb52952a79 req-525f6c1f-6837-4e64-afdd-36fb58cbf4d5 service nova] Acquiring lock "refresh_cache-1c7e408a-392d-4fcf-bae8-588ed9f92fec" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.060947] env[62552]: DEBUG oslo_concurrency.lockutils [req-7e537c6e-fee1-4656-949d-f6fb52952a79 req-525f6c1f-6837-4e64-afdd-36fb58cbf4d5 service nova] Acquired lock "refresh_cache-1c7e408a-392d-4fcf-bae8-588ed9f92fec" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.061367] env[62552]: DEBUG nova.network.neutron [req-7e537c6e-fee1-4656-949d-f6fb52952a79 req-525f6c1f-6837-4e64-afdd-36fb58cbf4d5 service nova] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Refreshing network info cache for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 761.073231] env[62552]: DEBUG nova.policy [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8df48ba6c0f2489f97397f25a73c5cc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d1a4de166344672a21e2b88f86668f7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 761.368036] env[62552]: ERROR nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614, please check neutron logs for more information. [ 761.368036] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.368036] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 761.368036] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.368036] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.368036] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.368036] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.368036] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.368036] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.368036] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 761.368036] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.368036] env[62552]: ERROR nova.compute.manager raise self.value [ 761.368036] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.368036] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.368036] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.368036] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.368628] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.368628] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.368628] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614, please check neutron logs for more information. [ 761.368628] env[62552]: ERROR nova.compute.manager [ 761.368628] env[62552]: Traceback (most recent call last): [ 761.368628] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.368628] env[62552]: listener.cb(fileno) [ 761.368628] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.368628] env[62552]: result = function(*args, **kwargs) [ 761.368628] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.368628] env[62552]: return func(*args, **kwargs) [ 761.368628] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 761.368628] env[62552]: raise e [ 761.368628] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 761.368628] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 761.368628] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.368628] env[62552]: created_port_ids = self._update_ports_for_instance( [ 761.368628] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.368628] env[62552]: with excutils.save_and_reraise_exception(): [ 761.368628] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.368628] env[62552]: self.force_reraise() [ 761.368628] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.368628] env[62552]: raise self.value [ 761.368628] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.368628] env[62552]: updated_port = self._update_port( [ 761.368628] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.368628] env[62552]: _ensure_no_port_binding_failure(port) [ 761.368628] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.368628] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.369449] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614, please check neutron logs for more information. [ 761.369449] env[62552]: Removing descriptor: 16 [ 761.369449] env[62552]: ERROR nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614, please check neutron logs for more information. [ 761.369449] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Traceback (most recent call last): [ 761.369449] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 761.369449] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] yield resources [ 761.369449] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 761.369449] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self.driver.spawn(context, instance, image_meta, [ 761.369449] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 761.369449] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.369449] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.369449] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] vm_ref = self.build_virtual_machine(instance, [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] for vif in network_info: [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] return self._sync_wrapper(fn, *args, **kwargs) [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self.wait() [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self[:] = self._gt.wait() [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] return self._exit_event.wait() [ 761.369804] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] result = hub.switch() [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] return self.greenlet.switch() [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] result = function(*args, **kwargs) [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] return func(*args, **kwargs) [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] raise e [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] nwinfo = self.network_api.allocate_for_instance( [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.370304] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] created_port_ids = self._update_ports_for_instance( [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] with excutils.save_and_reraise_exception(): [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self.force_reraise() [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] raise self.value [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] updated_port = self._update_port( [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] _ensure_no_port_binding_failure(port) [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.370685] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] raise exception.PortBindingFailed(port_id=port['id']) [ 761.371046] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] nova.exception.PortBindingFailed: Binding failed for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614, please check neutron logs for more information. [ 761.371046] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] [ 761.371046] env[62552]: INFO nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Terminating instance [ 761.439618] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Releasing lock "refresh_cache-ab363b52-ade2-45c4-ba3a-ae46ffaf831e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.440133] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 761.440343] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 761.440625] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b302b42b-f1c0-43e5-bbe5-47e059c432f7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.449930] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bced238e-ec5e-41cc-ab51-05dbaafe48ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.471684] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab363b52-ade2-45c4-ba3a-ae46ffaf831e could not be found. [ 761.471872] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.472098] env[62552]: INFO nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 761.472373] env[62552]: DEBUG oslo.service.loopingcall [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.472607] env[62552]: DEBUG nova.compute.manager [-] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 761.472706] env[62552]: DEBUG nova.network.neutron [-] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.495240] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 761.505409] env[62552]: DEBUG nova.network.neutron [-] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.599379] env[62552]: DEBUG nova.network.neutron [req-7e537c6e-fee1-4656-949d-f6fb52952a79 req-525f6c1f-6837-4e64-afdd-36fb58cbf4d5 service nova] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.601913] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Successfully created port: 4eb4b194-659d-40f4-b9fb-f110b64965fa {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 761.800890] env[62552]: DEBUG nova.network.neutron [req-7e537c6e-fee1-4656-949d-f6fb52952a79 req-525f6c1f-6837-4e64-afdd-36fb58cbf4d5 service nova] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.875470] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Acquiring lock "refresh_cache-1c7e408a-392d-4fcf-bae8-588ed9f92fec" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.898019] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a059122-c8bc-4a73-93fd-26c117d563e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.904335] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4764edc-9f7e-4af1-ad75-f8f951851f62 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.935476] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea07685-7476-4ddf-ab72-1f96175d1a83 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.943501] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b66f6f9-296b-4c62-b6a5-da04ff5fe350 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.956553] env[62552]: DEBUG nova.compute.provider_tree [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.010306] env[62552]: DEBUG nova.network.neutron [-] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.303668] env[62552]: DEBUG oslo_concurrency.lockutils [req-7e537c6e-fee1-4656-949d-f6fb52952a79 req-525f6c1f-6837-4e64-afdd-36fb58cbf4d5 service nova] Releasing lock "refresh_cache-1c7e408a-392d-4fcf-bae8-588ed9f92fec" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.304120] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Acquired lock "refresh_cache-1c7e408a-392d-4fcf-bae8-588ed9f92fec" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.304374] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.459981] env[62552]: DEBUG nova.scheduler.client.report [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 762.515589] env[62552]: INFO nova.compute.manager [-] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Took 1.04 seconds to deallocate network for instance. [ 762.516855] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 762.520974] env[62552]: DEBUG nova.compute.claims [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 762.521188] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.552805] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.553082] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.553257] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.553437] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.553575] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.553713] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.554012] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.554118] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.554316] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.554491] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.554673] env[62552]: DEBUG nova.virt.hardware [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.556850] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c0197f-1f8a-49e9-b8f3-43ec0c933a11 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.563858] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5254b8-1239-4c72-b5c5-3d377563d77b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.813117] env[62552]: ERROR nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4eb4b194-659d-40f4-b9fb-f110b64965fa, please check neutron logs for more information. [ 762.813117] env[62552]: ERROR nova.compute.manager Traceback (most recent call last): [ 762.813117] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 762.813117] env[62552]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 762.813117] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.813117] env[62552]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 762.813117] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.813117] env[62552]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 762.813117] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.813117] env[62552]: ERROR nova.compute.manager self.force_reraise() [ 762.813117] env[62552]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.813117] env[62552]: ERROR nova.compute.manager raise self.value [ 762.813117] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.813117] env[62552]: ERROR nova.compute.manager updated_port = self._update_port( [ 762.813117] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.813117] env[62552]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 762.813854] env[62552]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.813854] env[62552]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 762.813854] env[62552]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4eb4b194-659d-40f4-b9fb-f110b64965fa, please check neutron logs for more information. [ 762.813854] env[62552]: ERROR nova.compute.manager [ 762.813854] env[62552]: Traceback (most recent call last): [ 762.813854] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 762.813854] env[62552]: listener.cb(fileno) [ 762.813854] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.813854] env[62552]: result = function(*args, **kwargs) [ 762.813854] env[62552]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 762.813854] env[62552]: return func(*args, **kwargs) [ 762.813854] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 762.813854] env[62552]: raise e [ 762.813854] env[62552]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 762.813854] env[62552]: nwinfo = self.network_api.allocate_for_instance( [ 762.813854] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.813854] env[62552]: created_port_ids = self._update_ports_for_instance( [ 762.813854] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.813854] env[62552]: with excutils.save_and_reraise_exception(): [ 762.813854] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.813854] env[62552]: self.force_reraise() [ 762.813854] env[62552]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.813854] env[62552]: raise self.value [ 762.813854] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.813854] env[62552]: updated_port = self._update_port( [ 762.813854] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.813854] env[62552]: _ensure_no_port_binding_failure(port) [ 762.813854] env[62552]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.813854] env[62552]: raise exception.PortBindingFailed(port_id=port['id']) [ 762.814780] env[62552]: nova.exception.PortBindingFailed: Binding failed for port 4eb4b194-659d-40f4-b9fb-f110b64965fa, please check neutron logs for more information. [ 762.814780] env[62552]: Removing descriptor: 15 [ 762.814780] env[62552]: ERROR nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4eb4b194-659d-40f4-b9fb-f110b64965fa, please check neutron logs for more information. [ 762.814780] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Traceback (most recent call last): [ 762.814780] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 762.814780] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] yield resources [ 762.814780] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 762.814780] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self.driver.spawn(context, instance, image_meta, [ 762.814780] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 762.814780] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.814780] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.814780] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] vm_ref = self.build_virtual_machine(instance, [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] for vif in network_info: [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] return self._sync_wrapper(fn, *args, **kwargs) [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self.wait() [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self[:] = self._gt.wait() [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] return self._exit_event.wait() [ 762.815182] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] result = hub.switch() [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] return self.greenlet.switch() [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] result = function(*args, **kwargs) [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] return func(*args, **kwargs) [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] raise e [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] nwinfo = self.network_api.allocate_for_instance( [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.815606] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] created_port_ids = self._update_ports_for_instance( [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] with excutils.save_and_reraise_exception(): [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self.force_reraise() [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] raise self.value [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] updated_port = self._update_port( [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] _ensure_no_port_binding_failure(port) [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.816040] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] raise exception.PortBindingFailed(port_id=port['id']) [ 762.816437] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] nova.exception.PortBindingFailed: Binding failed for port 4eb4b194-659d-40f4-b9fb-f110b64965fa, please check neutron logs for more information. [ 762.816437] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] [ 762.816437] env[62552]: INFO nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Terminating instance [ 762.828762] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.929837] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.964486] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.964959] env[62552]: DEBUG nova.compute.manager [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 762.967553] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.315s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.968945] env[62552]: INFO nova.compute.claims [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.124144] env[62552]: DEBUG nova.compute.manager [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Received event network-vif-deleted-5e01d6b7-47ac-44d6-bf2f-097cb2207614 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 763.124358] env[62552]: DEBUG nova.compute.manager [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Received event network-changed-4eb4b194-659d-40f4-b9fb-f110b64965fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 763.124568] env[62552]: DEBUG nova.compute.manager [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Refreshing instance network info cache due to event network-changed-4eb4b194-659d-40f4-b9fb-f110b64965fa. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 763.124790] env[62552]: DEBUG oslo_concurrency.lockutils [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] Acquiring lock "refresh_cache-1252f45a-d55d-44c0-8629-aba51fb89956" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.124930] env[62552]: DEBUG oslo_concurrency.lockutils [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] Acquired lock "refresh_cache-1252f45a-d55d-44c0-8629-aba51fb89956" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.125099] env[62552]: DEBUG nova.network.neutron [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Refreshing network info cache for port 4eb4b194-659d-40f4-b9fb-f110b64965fa {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 763.319025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Acquiring lock "refresh_cache-1252f45a-d55d-44c0-8629-aba51fb89956" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.432387] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Releasing lock "refresh_cache-1c7e408a-392d-4fcf-bae8-588ed9f92fec" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.432888] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 763.433092] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 763.433388] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20a9be26-8672-4d29-8588-7a1c1bbb071c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.442848] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95094fb-7e34-4c2e-bc94-d33632099fea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.469249] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1c7e408a-392d-4fcf-bae8-588ed9f92fec could not be found. [ 763.469535] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.469725] env[62552]: INFO nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Took 0.04 seconds to destroy the instance on the hypervisor. [ 763.469997] env[62552]: DEBUG oslo.service.loopingcall [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.470260] env[62552]: DEBUG nova.compute.manager [-] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 763.470357] env[62552]: DEBUG nova.network.neutron [-] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.476351] env[62552]: DEBUG nova.compute.utils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.479497] env[62552]: DEBUG nova.compute.manager [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 763.479728] env[62552]: DEBUG nova.network.neutron [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 763.502400] env[62552]: DEBUG nova.network.neutron [-] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.534754] env[62552]: DEBUG nova.policy [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f87fac620384ff28a187546d6d29e2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25030025e90243b290a7d90efa26ec79', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.647297] env[62552]: DEBUG nova.network.neutron [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.847751] env[62552]: DEBUG nova.network.neutron [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.980827] env[62552]: DEBUG nova.compute.manager [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 764.015234] env[62552]: DEBUG nova.network.neutron [-] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.057081] env[62552]: DEBUG nova.network.neutron [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Successfully created port: 929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.327538] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762b34de-7a2f-470c-8b84-10ac78330713 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.335265] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8236ebbc-dc43-4d11-bf10-bbc8357b005f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.366057] env[62552]: DEBUG oslo_concurrency.lockutils [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] Releasing lock "refresh_cache-1252f45a-d55d-44c0-8629-aba51fb89956" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.366331] env[62552]: DEBUG nova.compute.manager [req-ba2b1a38-064e-4406-9388-0e6920f22a18 req-40950de7-78d6-4262-a6bb-aca1e3d5ae23 service nova] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Received event network-vif-deleted-4eb4b194-659d-40f4-b9fb-f110b64965fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 764.366879] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Acquired lock "refresh_cache-1252f45a-d55d-44c0-8629-aba51fb89956" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.367065] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.368587] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763108c8-5488-4278-b076-0a90c1d5e155 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.377900] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359dec4e-cb45-4aef-a5c6-72f385a6ec4e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.390177] env[62552]: DEBUG nova.compute.provider_tree [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.521032] env[62552]: INFO nova.compute.manager [-] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Took 1.05 seconds to deallocate network for instance. [ 764.523099] env[62552]: DEBUG nova.compute.claims [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 764.523314] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.888187] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.893464] env[62552]: DEBUG nova.scheduler.client.report [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 764.999233] env[62552]: DEBUG nova.compute.manager [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 765.008191] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.031347] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.031605] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.031760] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.031937] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.032377] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.032560] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.032771] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.032929] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.033108] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.033597] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.033597] env[62552]: DEBUG nova.virt.hardware [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.034558] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78768dea-bac8-442a-9be5-203dc70f3110 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.044103] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611b1daf-17ff-462f-86a7-9bd91d90d9ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.397500] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.398189] env[62552]: DEBUG nova.compute.manager [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 765.400594] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.859s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.402038] env[62552]: INFO nova.compute.claims [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.512630] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Releasing lock "refresh_cache-1252f45a-d55d-44c0-8629-aba51fb89956" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.513151] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 765.513294] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 765.513581] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f5d31e2-0d46-4c2f-aeed-59a8864a727a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.522469] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39dbb9c5-1497-4780-9ead-66a663a0fc55 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.544391] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1252f45a-d55d-44c0-8629-aba51fb89956 could not be found. [ 765.544391] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 765.544526] env[62552]: INFO nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Took 0.03 seconds to destroy the instance on the hypervisor. [ 765.545171] env[62552]: DEBUG oslo.service.loopingcall [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.545171] env[62552]: DEBUG nova.compute.manager [-] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 765.545171] env[62552]: DEBUG nova.network.neutron [-] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.579062] env[62552]: DEBUG nova.network.neutron [-] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.611659] env[62552]: DEBUG nova.compute.manager [req-4b11d6e5-221c-4e73-94f0-542fd31ca6aa req-58dbe191-5563-49f6-b77b-ec12b5170f90 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received event network-vif-plugged-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 765.611882] env[62552]: DEBUG oslo_concurrency.lockutils [req-4b11d6e5-221c-4e73-94f0-542fd31ca6aa req-58dbe191-5563-49f6-b77b-ec12b5170f90 service nova] Acquiring lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.612955] env[62552]: DEBUG oslo_concurrency.lockutils [req-4b11d6e5-221c-4e73-94f0-542fd31ca6aa req-58dbe191-5563-49f6-b77b-ec12b5170f90 service nova] Lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.612955] env[62552]: DEBUG oslo_concurrency.lockutils [req-4b11d6e5-221c-4e73-94f0-542fd31ca6aa req-58dbe191-5563-49f6-b77b-ec12b5170f90 service nova] Lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.612955] env[62552]: DEBUG nova.compute.manager [req-4b11d6e5-221c-4e73-94f0-542fd31ca6aa req-58dbe191-5563-49f6-b77b-ec12b5170f90 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] No waiting events found dispatching network-vif-plugged-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 765.612955] env[62552]: WARNING nova.compute.manager [req-4b11d6e5-221c-4e73-94f0-542fd31ca6aa req-58dbe191-5563-49f6-b77b-ec12b5170f90 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received unexpected event network-vif-plugged-929d55b0-5a35-4d2d-a172-15d79a215977 for instance with vm_state building and task_state spawning. [ 765.686705] env[62552]: DEBUG nova.network.neutron [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Successfully updated port: 929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 765.906369] env[62552]: DEBUG nova.compute.utils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.909888] env[62552]: DEBUG nova.compute.manager [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Not allocating networking since 'none' was specified. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 766.082058] env[62552]: DEBUG nova.network.neutron [-] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.189908] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.190141] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.190305] env[62552]: DEBUG nova.network.neutron [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.410729] env[62552]: DEBUG nova.compute.manager [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 766.585698] env[62552]: INFO nova.compute.manager [-] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Took 1.04 seconds to deallocate network for instance. [ 766.588030] env[62552]: DEBUG nova.compute.claims [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Aborting claim: {{(pid=62552) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 766.588236] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.704406] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382268f0-20f3-4c2d-9396-437347a0f1bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.712067] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c15bbd-1a90-42a6-b8d8-ad084071a3ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.742011] env[62552]: DEBUG nova.network.neutron [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.744223] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0daa358-24f8-4d12-9ead-b4f6b48e48ec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.751577] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3cc60a-2524-406d-95bb-c7c3261319eb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.764270] env[62552]: DEBUG nova.compute.provider_tree [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.969320] env[62552]: DEBUG nova.network.neutron [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating instance_info_cache with network_info: [{"id": "929d55b0-5a35-4d2d-a172-15d79a215977", "address": "fa:16:3e:62:4d:9b", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap929d55b0-5a", "ovs_interfaceid": "929d55b0-5a35-4d2d-a172-15d79a215977", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.267061] env[62552]: DEBUG nova.scheduler.client.report [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 767.360341] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.360341] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.424711] env[62552]: DEBUG nova.compute.manager [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 767.450982] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.451172] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.451275] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.451516] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.451597] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.451738] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.451945] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.452149] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.452276] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.452437] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.452607] env[62552]: DEBUG nova.virt.hardware [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.453486] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436672e1-6215-45a4-8d06-40bf6df4732f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.461859] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84bbb51-c7a5-4fe6-86e3-c49fade2f20c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.475328] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.475692] env[62552]: DEBUG nova.compute.manager [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Instance network_info: |[{"id": "929d55b0-5a35-4d2d-a172-15d79a215977", "address": "fa:16:3e:62:4d:9b", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap929d55b0-5a", "ovs_interfaceid": "929d55b0-5a35-4d2d-a172-15d79a215977", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 767.476143] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.482167] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Creating folder: Project (8f4030c4ff8f4f80af6c1871c8b159c8). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.482619] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:4d:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '929d55b0-5a35-4d2d-a172-15d79a215977', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.489770] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating folder: Project (25030025e90243b290a7d90efa26ec79). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.490012] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65ad6694-4397-46e2-bb9b-e73ccf315a18 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.491629] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1da71614-28b5-4adb-be98-ccb28722020e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.501648] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Created folder: Project (8f4030c4ff8f4f80af6c1871c8b159c8) in parent group-v267339. [ 767.501819] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Creating folder: Instances. Parent ref: group-v267360. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.502066] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4963d990-5750-4171-afde-f4c35644c4f5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.504408] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Created folder: Project (25030025e90243b290a7d90efa26ec79) in parent group-v267339. [ 767.504578] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating folder: Instances. Parent ref: group-v267361. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.505065] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d6c6aab7-702e-4cb8-919b-2f161356e798 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.512250] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Created folder: Instances in parent group-v267360. [ 767.512461] env[62552]: DEBUG oslo.service.loopingcall [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.512632] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 767.512808] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84c7b802-a874-4d6a-b669-e08b30200515 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.524830] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Created folder: Instances in parent group-v267361. [ 767.525049] env[62552]: DEBUG oslo.service.loopingcall [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.525488] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 767.525670] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a318d638-f4a9-42ec-b955-abf3b2422cc0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.539654] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.539654] env[62552]: value = "task-1239260" [ 767.539654] env[62552]: _type = "Task" [ 767.539654] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.543968] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.543968] env[62552]: value = "task-1239261" [ 767.543968] env[62552]: _type = "Task" [ 767.543968] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.549356] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239260, 'name': CreateVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.554123] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239261, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.645670] env[62552]: DEBUG nova.compute.manager [req-5bb312db-4fc0-4758-a4aa-b804c66f58fe req-1fc73aff-c499-4e96-a857-a2246762d76e service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received event network-changed-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 767.646688] env[62552]: DEBUG nova.compute.manager [req-5bb312db-4fc0-4758-a4aa-b804c66f58fe req-1fc73aff-c499-4e96-a857-a2246762d76e service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Refreshing instance network info cache due to event network-changed-929d55b0-5a35-4d2d-a172-15d79a215977. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 767.646688] env[62552]: DEBUG oslo_concurrency.lockutils [req-5bb312db-4fc0-4758-a4aa-b804c66f58fe req-1fc73aff-c499-4e96-a857-a2246762d76e service nova] Acquiring lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.646688] env[62552]: DEBUG oslo_concurrency.lockutils [req-5bb312db-4fc0-4758-a4aa-b804c66f58fe req-1fc73aff-c499-4e96-a857-a2246762d76e service nova] Acquired lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.646688] env[62552]: DEBUG nova.network.neutron [req-5bb312db-4fc0-4758-a4aa-b804c66f58fe req-1fc73aff-c499-4e96-a857-a2246762d76e service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Refreshing network info cache for port 929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 767.772041] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.772442] env[62552]: DEBUG nova.compute.manager [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 767.775398] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.894s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.864919] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.864919] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 767.865157] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Rebuilding the list of instances to heal {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 768.054997] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239260, 'name': CreateVM_Task, 'duration_secs': 0.242004} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.056771] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 768.057080] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239261, 'name': CreateVM_Task, 'duration_secs': 0.295592} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.057644] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.057898] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.058333] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 768.058519] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 768.058724] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec0cee68-33d2-482e-bf83-61c03c03df8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.064265] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 768.064265] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a81417-c57e-2b21-84c1-224e8295662e" [ 768.064265] env[62552]: _type = "Task" [ 768.064265] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.071882] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a81417-c57e-2b21-84c1-224e8295662e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.072776] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.280448] env[62552]: DEBUG nova.compute.utils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.289657] env[62552]: DEBUG nova.compute.manager [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 768.289840] env[62552]: DEBUG nova.network.neutron [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 768.368608] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 768.368770] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 768.368896] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 768.369032] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 768.369182] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 768.369327] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 768.369448] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Didn't find any instances for network info cache update. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10357}} [ 768.370908] env[62552]: DEBUG nova.policy [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbd041c5525240639d057e14a3875294', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '56e85e5029e347e89447147bfc0ee79b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 768.373117] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.373276] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.376021] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.376364] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.376528] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.376678] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.376803] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 768.376941] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.575317] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a81417-c57e-2b21-84c1-224e8295662e, 'name': SearchDatastore_Task, 'duration_secs': 0.009172} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.575638] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.575869] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 768.576205] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.576366] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.576550] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 768.577150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.577647] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 768.577729] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53fb94ae-8ca5-43dd-80c1-9fb87e5fede4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.579494] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09d48ba8-b1c0-4a4d-84d3-efa8e41404f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.585863] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 768.585863] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f4ebdb-4a15-8361-d6a0-1ac783d636b5" [ 768.585863] env[62552]: _type = "Task" [ 768.585863] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.586570] env[62552]: DEBUG nova.network.neutron [req-5bb312db-4fc0-4758-a4aa-b804c66f58fe req-1fc73aff-c499-4e96-a857-a2246762d76e service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updated VIF entry in instance network info cache for port 929d55b0-5a35-4d2d-a172-15d79a215977. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 768.586933] env[62552]: DEBUG nova.network.neutron [req-5bb312db-4fc0-4758-a4aa-b804c66f58fe req-1fc73aff-c499-4e96-a857-a2246762d76e service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating instance_info_cache with network_info: [{"id": "929d55b0-5a35-4d2d-a172-15d79a215977", "address": "fa:16:3e:62:4d:9b", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap929d55b0-5a", "ovs_interfaceid": "929d55b0-5a35-4d2d-a172-15d79a215977", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.593768] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 768.593941] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 768.594981] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-150c9277-d52f-4f09-a19a-b0392f7457f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.601389] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f4ebdb-4a15-8361-d6a0-1ac783d636b5, 'name': SearchDatastore_Task, 'duration_secs': 0.00851} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.602663] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.602911] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 768.603144] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.608033] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 768.608033] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524a805d-36a5-a1ad-2a0e-eddbacb6e83a" [ 768.608033] env[62552]: _type = "Task" [ 768.608033] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.616696] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524a805d-36a5-a1ad-2a0e-eddbacb6e83a, 'name': SearchDatastore_Task, 'duration_secs': 0.008514} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.619674] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9902660f-5678-4fae-b14f-4dfac2dce1ed {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.624578] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 768.624578] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52593890-196d-d7b6-3670-ec0a3fd5ca93" [ 768.624578] env[62552]: _type = "Task" [ 768.624578] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.630881] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b20f1a-45ca-4c31-bca1-1707d64702f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.637814] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52593890-196d-d7b6-3670-ec0a3fd5ca93, 'name': SearchDatastore_Task, 'duration_secs': 0.007851} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.639959] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.640265] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282/1472ca01-ba04-4e7d-a9cf-ccaf29c04282.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 768.640559] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.640739] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 768.640944] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a671def-e5b7-42a0-84d7-2f286935f54b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.643543] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1650f30-e98e-4ffe-9b40-675eb86bee06 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.646858] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d99d965-0dbe-4fc8-b7e1-3e090b9bec4a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.654131] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 768.654131] env[62552]: value = "task-1239262" [ 768.654131] env[62552]: _type = "Task" [ 768.654131] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.683454] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5e8498-8c22-42ac-9df9-b207bc4d7943 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.694446] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d6efd6-5451-4e08-9b2c-b150a5239c30 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.698429] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.699512] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 768.699702] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 768.700738] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64dfad7d-94ea-41b2-b8ee-cb2857eab4d1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.711454] env[62552]: DEBUG nova.compute.provider_tree [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.715178] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 768.715178] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a7aed5-31ba-d329-225f-5161f0cdf1a0" [ 768.715178] env[62552]: _type = "Task" [ 768.715178] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.722832] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a7aed5-31ba-d329-225f-5161f0cdf1a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.787599] env[62552]: DEBUG nova.compute.manager [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 768.880575] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.882575] env[62552]: DEBUG nova.network.neutron [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Successfully created port: 65c6abb5-29b3-4603-b019-4716db17392c {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.093995] env[62552]: DEBUG oslo_concurrency.lockutils [req-5bb312db-4fc0-4758-a4aa-b804c66f58fe req-1fc73aff-c499-4e96-a857-a2246762d76e service nova] Releasing lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.188787] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239262, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435459} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.189336] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282/1472ca01-ba04-4e7d-a9cf-ccaf29c04282.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 769.189718] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.190094] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e8427f9-b07e-464e-8145-4b92946cd8d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.197021] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 769.197021] env[62552]: value = "task-1239263" [ 769.197021] env[62552]: _type = "Task" [ 769.197021] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.206226] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239263, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.214242] env[62552]: DEBUG nova.scheduler.client.report [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 769.227137] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a7aed5-31ba-d329-225f-5161f0cdf1a0, 'name': SearchDatastore_Task, 'duration_secs': 0.006767} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.228053] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc624ac6-b07d-45e6-8c69-7c798caa7326 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.233617] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 769.233617] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c25cce-5946-1a98-bb02-8a0631baaac3" [ 769.233617] env[62552]: _type = "Task" [ 769.233617] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.241680] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c25cce-5946-1a98-bb02-8a0631baaac3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.294884] env[62552]: INFO nova.virt.block_device [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Booting with volume 0feb401f-606c-4b63-8c60-46d9d717225f at /dev/sda [ 769.340678] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de4d9602-7ddc-4b0d-831b-5ad7c52657fc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.349880] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db096761-7c68-4fb2-aa83-cb8378d227de {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.377762] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6f3ee80-25dd-4f1c-9fc9-8f1956b22758 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.385620] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2524469-2234-4054-865b-e8e8fa201f3f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.408395] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a875940e-fab2-481f-890f-3a7738feffab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.414388] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73874f8d-e31d-4d81-b1d8-305c3dbb4b2f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.427568] env[62552]: DEBUG nova.virt.block_device [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Updating existing volume attachment record: 48275610-1965-4177-bfd8-87d67ca70678 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 769.708137] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239263, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063834} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.708396] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 769.709171] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff58f565-4b8a-4224-bc53-ecbb3592d243 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.728497] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282/1472ca01-ba04-4e7d-a9cf-ccaf29c04282.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 769.729239] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.729831] env[62552]: ERROR nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 365eabca-4338-4379-8328-135cccda66fa, please check neutron logs for more information. [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Traceback (most recent call last): [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self.driver.spawn(context, instance, image_meta, [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] vm_ref = self.build_virtual_machine(instance, [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.729831] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] for vif in network_info: [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] return self._sync_wrapper(fn, *args, **kwargs) [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self.wait() [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self[:] = self._gt.wait() [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] return self._exit_event.wait() [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] result = hub.switch() [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 769.730346] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] return self.greenlet.switch() [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] result = function(*args, **kwargs) [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] return func(*args, **kwargs) [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] raise e [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] nwinfo = self.network_api.allocate_for_instance( [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] created_port_ids = self._update_ports_for_instance( [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] with excutils.save_and_reraise_exception(): [ 769.730707] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] self.force_reraise() [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] raise self.value [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] updated_port = self._update_port( [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] _ensure_no_port_binding_failure(port) [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] raise exception.PortBindingFailed(port_id=port['id']) [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] nova.exception.PortBindingFailed: Binding failed for port 365eabca-4338-4379-8328-135cccda66fa, please check neutron logs for more information. [ 769.731110] env[62552]: ERROR nova.compute.manager [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] [ 769.731418] env[62552]: DEBUG nova.compute.utils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Binding failed for port 365eabca-4338-4379-8328-135cccda66fa, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.731617] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-171ec247-7f1b-43dc-9a1b-3bb059913a9f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.745590] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Build of instance e8ec980c-0186-48cb-84b9-af902f23d2fb was re-scheduled: Binding failed for port 365eabca-4338-4379-8328-135cccda66fa, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 769.746059] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 769.746297] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Acquiring lock "refresh_cache-e8ec980c-0186-48cb-84b9-af902f23d2fb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.746498] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Acquired lock "refresh_cache-e8ec980c-0186-48cb-84b9-af902f23d2fb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.746621] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.747633] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.875s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.749537] env[62552]: INFO nova.compute.claims [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.762847] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c25cce-5946-1a98-bb02-8a0631baaac3, 'name': SearchDatastore_Task, 'duration_secs': 0.007999} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.764244] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.764640] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] cdee22a9-4327-47af-ab14-8403a06b2802/cdee22a9-4327-47af-ab14-8403a06b2802.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 769.764828] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 769.764828] env[62552]: value = "task-1239264" [ 769.764828] env[62552]: _type = "Task" [ 769.764828] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.765056] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7596362-8b8b-4e29-a520-7b48e11a3442 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.778361] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 769.778361] env[62552]: value = "task-1239265" [ 769.778361] env[62552]: _type = "Task" [ 769.778361] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.779191] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239264, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.787694] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239265, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.273639] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.281708] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239264, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.290668] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239265, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504142} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.290889] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] cdee22a9-4327-47af-ab14-8403a06b2802/cdee22a9-4327-47af-ab14-8403a06b2802.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 770.292448] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 770.292448] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21dc48c5-5a28-446d-b533-289d5944e859 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.298530] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 770.298530] env[62552]: value = "task-1239266" [ 770.298530] env[62552]: _type = "Task" [ 770.298530] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.306912] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239266, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.397558] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.780028] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239264, 'name': ReconfigVM_Task, 'duration_secs': 0.573843} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.780028] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282/1472ca01-ba04-4e7d-a9cf-ccaf29c04282.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.780391] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e91fbe3-f402-4e8e-b831-8fcd046c5894 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.791857] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 770.791857] env[62552]: value = "task-1239267" [ 770.791857] env[62552]: _type = "Task" [ 770.791857] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.798921] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239267, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.810782] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239266, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085086} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.810782] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 770.812687] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0610222c-29ac-4e1e-8592-d5ac9955c990 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.837745] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] cdee22a9-4327-47af-ab14-8403a06b2802/cdee22a9-4327-47af-ab14-8403a06b2802.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 770.838409] env[62552]: DEBUG nova.network.neutron [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Successfully updated port: 65c6abb5-29b3-4603-b019-4716db17392c {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 770.842992] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-358b9c60-d0ff-43d8-83c2-d445ef2c8f27 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.861862] env[62552]: DEBUG nova.compute.manager [req-6575c465-4ac3-46cc-a04d-bef4986e3700 req-9181a337-d0bc-4c22-a0f5-78c9d614ef81 service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Received event network-vif-plugged-65c6abb5-29b3-4603-b019-4716db17392c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 770.861862] env[62552]: DEBUG oslo_concurrency.lockutils [req-6575c465-4ac3-46cc-a04d-bef4986e3700 req-9181a337-d0bc-4c22-a0f5-78c9d614ef81 service nova] Acquiring lock "8707ce4b-677e-4f13-86f8-3e327d19380b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.862630] env[62552]: DEBUG oslo_concurrency.lockutils [req-6575c465-4ac3-46cc-a04d-bef4986e3700 req-9181a337-d0bc-4c22-a0f5-78c9d614ef81 service nova] Lock "8707ce4b-677e-4f13-86f8-3e327d19380b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.862630] env[62552]: DEBUG oslo_concurrency.lockutils [req-6575c465-4ac3-46cc-a04d-bef4986e3700 req-9181a337-d0bc-4c22-a0f5-78c9d614ef81 service nova] Lock "8707ce4b-677e-4f13-86f8-3e327d19380b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.862630] env[62552]: DEBUG nova.compute.manager [req-6575c465-4ac3-46cc-a04d-bef4986e3700 req-9181a337-d0bc-4c22-a0f5-78c9d614ef81 service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] No waiting events found dispatching network-vif-plugged-65c6abb5-29b3-4603-b019-4716db17392c {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 770.862630] env[62552]: WARNING nova.compute.manager [req-6575c465-4ac3-46cc-a04d-bef4986e3700 req-9181a337-d0bc-4c22-a0f5-78c9d614ef81 service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Received unexpected event network-vif-plugged-65c6abb5-29b3-4603-b019-4716db17392c for instance with vm_state building and task_state block_device_mapping. [ 770.868028] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 770.868028] env[62552]: value = "task-1239268" [ 770.868028] env[62552]: _type = "Task" [ 770.868028] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.880485] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239268, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.900812] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Releasing lock "refresh_cache-e8ec980c-0186-48cb-84b9-af902f23d2fb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.901119] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 770.901385] env[62552]: DEBUG nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 770.901559] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.928667] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.177479] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d483621b-e983-46c8-a552-6626db362638 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.184742] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4dcdb1f-a387-484a-8da2-ae82755b36c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.216012] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ab49ff-a53e-4d26-9e20-430963e307e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.223361] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd634705-684b-4271-a007-4530a0e8e149 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.236082] env[62552]: DEBUG nova.compute.provider_tree [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.298819] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239267, 'name': Rename_Task, 'duration_secs': 0.135476} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.299145] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 771.299295] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc7df7dc-1a89-4e7c-a14d-c3522607bf25 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.305131] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 771.305131] env[62552]: value = "task-1239269" [ 771.305131] env[62552]: _type = "Task" [ 771.305131] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.312074] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239269, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.361065] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Acquiring lock "refresh_cache-8707ce4b-677e-4f13-86f8-3e327d19380b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.361249] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Acquired lock "refresh_cache-8707ce4b-677e-4f13-86f8-3e327d19380b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.361405] env[62552]: DEBUG nova.network.neutron [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.377466] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239268, 'name': ReconfigVM_Task, 'duration_secs': 0.285601} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.377787] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Reconfigured VM instance instance-0000002d to attach disk [datastore2] cdee22a9-4327-47af-ab14-8403a06b2802/cdee22a9-4327-47af-ab14-8403a06b2802.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 771.378440] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73301969-15d2-411c-81a5-8f4ba036951c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.383974] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 771.383974] env[62552]: value = "task-1239270" [ 771.383974] env[62552]: _type = "Task" [ 771.383974] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.392142] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239270, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.431093] env[62552]: DEBUG nova.network.neutron [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.539248] env[62552]: DEBUG nova.compute.manager [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 771.539835] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.539835] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.540041] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.540096] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.540483] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.540483] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.540632] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.540716] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.540923] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.542061] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.542061] env[62552]: DEBUG nova.virt.hardware [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.542061] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c89aae-c961-4359-834a-5d425cfc6298 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.550441] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1798fce-623b-43d3-ac3c-cc347ef1558c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.738810] env[62552]: DEBUG nova.scheduler.client.report [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 771.815483] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239269, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.895127] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239270, 'name': Rename_Task, 'duration_secs': 0.172837} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.895407] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 771.895642] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8903427-0986-4847-a1c1-9f9b06cbc62c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.902438] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 771.902438] env[62552]: value = "task-1239271" [ 771.902438] env[62552]: _type = "Task" [ 771.902438] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.910337] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239271, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.915220] env[62552]: DEBUG nova.network.neutron [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.933646] env[62552]: INFO nova.compute.manager [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] [instance: e8ec980c-0186-48cb-84b9-af902f23d2fb] Took 1.03 seconds to deallocate network for instance. [ 772.119117] env[62552]: DEBUG nova.network.neutron [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Updating instance_info_cache with network_info: [{"id": "65c6abb5-29b3-4603-b019-4716db17392c", "address": "fa:16:3e:2d:a8:73", "network": {"id": "7e24b87f-25ff-4556-a4df-5cf01e529ac5", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-395831708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "56e85e5029e347e89447147bfc0ee79b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c6abb5-29", "ovs_interfaceid": "65c6abb5-29b3-4603-b019-4716db17392c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.244240] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.244554] env[62552]: DEBUG nova.compute.manager [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 772.247252] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.424s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.249075] env[62552]: INFO nova.compute.claims [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.316038] env[62552]: DEBUG oslo_vmware.api [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239269, 'name': PowerOnVM_Task, 'duration_secs': 0.63373} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.316423] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 772.316503] env[62552]: INFO nova.compute.manager [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Took 4.89 seconds to spawn the instance on the hypervisor. [ 772.316679] env[62552]: DEBUG nova.compute.manager [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 772.317455] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be127576-b3c6-4b89-b1d9-12e0f033db5f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.412635] env[62552]: DEBUG oslo_vmware.api [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239271, 'name': PowerOnVM_Task, 'duration_secs': 0.471836} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.412906] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 772.413124] env[62552]: INFO nova.compute.manager [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Took 7.42 seconds to spawn the instance on the hypervisor. [ 772.413305] env[62552]: DEBUG nova.compute.manager [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 772.414042] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ba39c0-6fd9-4305-91c6-46e6e8fbf252 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.624666] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Releasing lock "refresh_cache-8707ce4b-677e-4f13-86f8-3e327d19380b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.625136] env[62552]: DEBUG nova.compute.manager [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Instance network_info: |[{"id": "65c6abb5-29b3-4603-b019-4716db17392c", "address": "fa:16:3e:2d:a8:73", "network": {"id": "7e24b87f-25ff-4556-a4df-5cf01e529ac5", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-395831708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "56e85e5029e347e89447147bfc0ee79b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c6abb5-29", "ovs_interfaceid": "65c6abb5-29b3-4603-b019-4716db17392c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 772.625699] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:a8:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b94712a6-b777-47dd-bc06-f9acfce2d936', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65c6abb5-29b3-4603-b019-4716db17392c', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 772.637758] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Creating folder: Project (56e85e5029e347e89447147bfc0ee79b). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 772.638268] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df014b5a-85f5-4e7b-82b8-1e700ea5257f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.651882] env[62552]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 772.652118] env[62552]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62552) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 772.652539] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Folder already exists: Project (56e85e5029e347e89447147bfc0ee79b). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 772.652814] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Creating folder: Instances. Parent ref: group-v267347. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 772.653124] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc8f4105-8023-48a5-9bab-2fd34b9c573d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.661383] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Created folder: Instances in parent group-v267347. [ 772.661599] env[62552]: DEBUG oslo.service.loopingcall [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.661775] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 772.661963] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bce36fa3-5852-4943-9677-39241e84d3b4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.679523] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 772.679523] env[62552]: value = "task-1239274" [ 772.679523] env[62552]: _type = "Task" [ 772.679523] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.686601] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239274, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.752794] env[62552]: DEBUG nova.compute.utils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.757353] env[62552]: DEBUG nova.compute.manager [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 772.757565] env[62552]: DEBUG nova.network.neutron [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 772.798430] env[62552]: DEBUG nova.policy [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '824c76bc94b044ed8151f99872e8e8d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2fe0e0c574e4ff2a2ed4acdd4071e42', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.838308] env[62552]: INFO nova.compute.manager [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Took 25.20 seconds to build instance. [ 772.935458] env[62552]: INFO nova.compute.manager [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Took 26.69 seconds to build instance. [ 772.972215] env[62552]: DEBUG nova.compute.manager [req-3a929d5b-81f3-4f1f-8cbc-8c46e617ef97 req-0ecd7872-dc88-4016-b345-3650a1d6205c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Received event network-changed-65c6abb5-29b3-4603-b019-4716db17392c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 772.972215] env[62552]: DEBUG nova.compute.manager [req-3a929d5b-81f3-4f1f-8cbc-8c46e617ef97 req-0ecd7872-dc88-4016-b345-3650a1d6205c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Refreshing instance network info cache due to event network-changed-65c6abb5-29b3-4603-b019-4716db17392c. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 772.972362] env[62552]: DEBUG oslo_concurrency.lockutils [req-3a929d5b-81f3-4f1f-8cbc-8c46e617ef97 req-0ecd7872-dc88-4016-b345-3650a1d6205c service nova] Acquiring lock "refresh_cache-8707ce4b-677e-4f13-86f8-3e327d19380b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.972603] env[62552]: DEBUG oslo_concurrency.lockutils [req-3a929d5b-81f3-4f1f-8cbc-8c46e617ef97 req-0ecd7872-dc88-4016-b345-3650a1d6205c service nova] Acquired lock "refresh_cache-8707ce4b-677e-4f13-86f8-3e327d19380b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.972682] env[62552]: DEBUG nova.network.neutron [req-3a929d5b-81f3-4f1f-8cbc-8c46e617ef97 req-0ecd7872-dc88-4016-b345-3650a1d6205c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Refreshing network info cache for port 65c6abb5-29b3-4603-b019-4716db17392c {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.975920] env[62552]: INFO nova.scheduler.client.report [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Deleted allocations for instance e8ec980c-0186-48cb-84b9-af902f23d2fb [ 773.196390] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239274, 'name': CreateVM_Task, 'duration_secs': 0.311408} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.196593] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 773.197316] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267350', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'name': 'volume-0feb401f-606c-4b63-8c60-46d9d717225f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8707ce4b-677e-4f13-86f8-3e327d19380b', 'attached_at': '', 'detached_at': '', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'serial': '0feb401f-606c-4b63-8c60-46d9d717225f'}, 'disk_bus': None, 'boot_index': 0, 'attachment_id': '48275610-1965-4177-bfd8-87d67ca70678', 'volume_type': None}], 'swap': None} {{(pid=62552) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 773.197521] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Root volume attach. Driver type: vmdk {{(pid=62552) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 773.198389] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eff26d0-9b08-4a95-be0c-e001eb4e8198 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.207175] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f166b8ff-c52c-4683-b248-7b8942b00a03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.213478] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fece294-bfad-4c6c-a077-721ba206c7d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.219407] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-8a8c1dd9-1ef0-4804-a3e0-4da232b63da9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.226195] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 773.226195] env[62552]: value = "task-1239275" [ 773.226195] env[62552]: _type = "Task" [ 773.226195] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.233589] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239275, 'name': RelocateVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.260954] env[62552]: DEBUG nova.compute.manager [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 773.335985] env[62552]: DEBUG nova.network.neutron [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Successfully created port: 5cb2de18-f6f3-4351-a156-891bb4a6c542 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.342969] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0ca81f22-2e31-4138-bd48-e75914da65a2 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "1472ca01-ba04-4e7d-a9cf-ccaf29c04282" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.996s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.438322] env[62552]: DEBUG oslo_concurrency.lockutils [None req-63ed2b3c-aa7f-4e72-909e-90d83014ee32 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.380s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.489482] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a4739829-17ac-49be-a18d-70ae2a50602a tempest-ServerActionsTestOtherB-2106020826 tempest-ServerActionsTestOtherB-2106020826-project-member] Lock "e8ec980c-0186-48cb-84b9-af902f23d2fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 170.410s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.639514] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986ba552-5f80-4d55-a00d-7beb721559d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.652788] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9d9853-5852-4363-9d95-946768de1e61 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.685386] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18a39cd-eff4-47b5-b55c-d7c9f1db3eb5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.699076] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b9f8a8-2a23-4246-a484-ba94321377cb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.713115] env[62552]: DEBUG nova.compute.provider_tree [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.735674] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239275, 'name': RelocateVM_Task, 'duration_secs': 0.355509} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.738331] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 773.738557] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267350', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'name': 'volume-0feb401f-606c-4b63-8c60-46d9d717225f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8707ce4b-677e-4f13-86f8-3e327d19380b', 'attached_at': '', 'detached_at': '', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'serial': '0feb401f-606c-4b63-8c60-46d9d717225f'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 773.739922] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb01b35c-866e-492e-9517-a9e816a0ea87 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.756558] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e7b751-8b4d-4579-8c44-2bd301db9dba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.765516] env[62552]: INFO nova.virt.block_device [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Booting with volume 1e3c30b0-8285-44ec-9795-8dca10fcf1ed at /dev/sda [ 773.785916] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] volume-0feb401f-606c-4b63-8c60-46d9d717225f/volume-0feb401f-606c-4b63-8c60-46d9d717225f.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 773.789610] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef82c4e3-5b01-4b5a-b57e-92bb4bf8219b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.810147] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 773.810147] env[62552]: value = "task-1239276" [ 773.810147] env[62552]: _type = "Task" [ 773.810147] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.818377] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239276, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.839910] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11d239c5-66a4-4496-a0d1-4505c66138ca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.842634] env[62552]: DEBUG nova.compute.manager [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 773.852970] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddac8c07-8806-410c-8f14-0e9cc439217d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.877421] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d96c294f-1870-421c-8c52-b73c428c7118 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.884766] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae65586f-1a99-4333-a17c-dc76368905d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.909134] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef3f65c-e913-491e-836b-2a28503fb783 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.916794] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0b299b-8ab6-44b2-ba09-7f4937c400d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.922401] env[62552]: DEBUG nova.network.neutron [req-3a929d5b-81f3-4f1f-8cbc-8c46e617ef97 req-0ecd7872-dc88-4016-b345-3650a1d6205c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Updated VIF entry in instance network info cache for port 65c6abb5-29b3-4603-b019-4716db17392c. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 773.922935] env[62552]: DEBUG nova.network.neutron [req-3a929d5b-81f3-4f1f-8cbc-8c46e617ef97 req-0ecd7872-dc88-4016-b345-3650a1d6205c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Updating instance_info_cache with network_info: [{"id": "65c6abb5-29b3-4603-b019-4716db17392c", "address": "fa:16:3e:2d:a8:73", "network": {"id": "7e24b87f-25ff-4556-a4df-5cf01e529ac5", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-395831708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "56e85e5029e347e89447147bfc0ee79b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c6abb5-29", "ovs_interfaceid": "65c6abb5-29b3-4603-b019-4716db17392c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.930335] env[62552]: DEBUG nova.virt.block_device [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Updating existing volume attachment record: 92799767-511b-4d29-af4c-507a783b74dc {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 773.941249] env[62552]: DEBUG nova.compute.manager [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 773.994343] env[62552]: DEBUG nova.compute.manager [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 774.216490] env[62552]: DEBUG nova.scheduler.client.report [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 774.264675] env[62552]: INFO nova.compute.manager [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Rebuilding instance [ 774.317964] env[62552]: DEBUG nova.compute.manager [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 774.319967] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf13dea-7e60-49e2-ad35-b6cced5390d3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.325282] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239276, 'name': ReconfigVM_Task, 'duration_secs': 0.241951} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.325878] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Reconfigured VM instance instance-0000002f to attach disk [datastore2] volume-0feb401f-606c-4b63-8c60-46d9d717225f/volume-0feb401f-606c-4b63-8c60-46d9d717225f.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 774.334106] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d290eb0-89d7-4146-ba5d-2c6930d34f16 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.357051] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 774.357051] env[62552]: value = "task-1239277" [ 774.357051] env[62552]: _type = "Task" [ 774.357051] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.368103] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239277, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.371780] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.425672] env[62552]: DEBUG oslo_concurrency.lockutils [req-3a929d5b-81f3-4f1f-8cbc-8c46e617ef97 req-0ecd7872-dc88-4016-b345-3650a1d6205c service nova] Releasing lock "refresh_cache-8707ce4b-677e-4f13-86f8-3e327d19380b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.454333] env[62552]: DEBUG nova.compute.manager [req-e0068a50-2c32-4cc6-a8e0-61f46f078948 req-7a537038-10e6-4330-8dc7-02b3f8e05f17 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received event network-changed-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 774.455202] env[62552]: DEBUG nova.compute.manager [req-e0068a50-2c32-4cc6-a8e0-61f46f078948 req-7a537038-10e6-4330-8dc7-02b3f8e05f17 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Refreshing instance network info cache due to event network-changed-929d55b0-5a35-4d2d-a172-15d79a215977. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 774.455202] env[62552]: DEBUG oslo_concurrency.lockutils [req-e0068a50-2c32-4cc6-a8e0-61f46f078948 req-7a537038-10e6-4330-8dc7-02b3f8e05f17 service nova] Acquiring lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.455202] env[62552]: DEBUG oslo_concurrency.lockutils [req-e0068a50-2c32-4cc6-a8e0-61f46f078948 req-7a537038-10e6-4330-8dc7-02b3f8e05f17 service nova] Acquired lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.455454] env[62552]: DEBUG nova.network.neutron [req-e0068a50-2c32-4cc6-a8e0-61f46f078948 req-7a537038-10e6-4330-8dc7-02b3f8e05f17 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Refreshing network info cache for port 929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 774.464351] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.511240] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.722431] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.722965] env[62552]: DEBUG nova.compute.manager [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 774.727149] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.891s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.728567] env[62552]: INFO nova.compute.claims [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.868071] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239277, 'name': ReconfigVM_Task, 'duration_secs': 0.121851} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.868433] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267350', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'name': 'volume-0feb401f-606c-4b63-8c60-46d9d717225f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8707ce4b-677e-4f13-86f8-3e327d19380b', 'attached_at': '', 'detached_at': '', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'serial': '0feb401f-606c-4b63-8c60-46d9d717225f'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 774.869030] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-06168da5-2439-44cd-868e-6277ef43b82e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.876447] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 774.876447] env[62552]: value = "task-1239278" [ 774.876447] env[62552]: _type = "Task" [ 774.876447] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.884658] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239278, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.237857] env[62552]: DEBUG nova.compute.utils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.239413] env[62552]: DEBUG nova.compute.manager [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 775.239578] env[62552]: DEBUG nova.network.neutron [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 775.352864] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 775.353157] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d28f7042-6ee1-4228-a6c6-bec8a59a4723 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.360680] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 775.360680] env[62552]: value = "task-1239279" [ 775.360680] env[62552]: _type = "Task" [ 775.360680] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.370086] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.386926] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239278, 'name': Rename_Task, 'duration_secs': 0.140253} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.388429] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 775.388691] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36ed033a-d70d-45bd-9c1f-ffb24472ec6b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.394787] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 775.394787] env[62552]: value = "task-1239280" [ 775.394787] env[62552]: _type = "Task" [ 775.394787] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.406425] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239280, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.468565] env[62552]: DEBUG nova.policy [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecfe0d4d977540d99709760db0564cf9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3758acde6e3e4320bf63d4cd1667bfc2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.746767] env[62552]: DEBUG nova.compute.manager [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 775.868639] env[62552]: DEBUG nova.network.neutron [req-e0068a50-2c32-4cc6-a8e0-61f46f078948 req-7a537038-10e6-4330-8dc7-02b3f8e05f17 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updated VIF entry in instance network info cache for port 929d55b0-5a35-4d2d-a172-15d79a215977. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 775.873344] env[62552]: DEBUG nova.network.neutron [req-e0068a50-2c32-4cc6-a8e0-61f46f078948 req-7a537038-10e6-4330-8dc7-02b3f8e05f17 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating instance_info_cache with network_info: [{"id": "929d55b0-5a35-4d2d-a172-15d79a215977", "address": "fa:16:3e:62:4d:9b", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap929d55b0-5a", "ovs_interfaceid": "929d55b0-5a35-4d2d-a172-15d79a215977", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.876196] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239279, 'name': PowerOffVM_Task, 'duration_secs': 0.33816} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.878892] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 775.878892] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.878892] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e9ae8b-c7ec-402b-ae49-84db01d6c6ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.891895] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 775.892927] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6c10db9-d36a-4cec-b875-779f47506fd6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.907291] env[62552]: DEBUG oslo_vmware.api [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239280, 'name': PowerOnVM_Task, 'duration_secs': 0.493242} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.907786] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 775.908594] env[62552]: INFO nova.compute.manager [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Took 4.37 seconds to spawn the instance on the hypervisor. [ 775.908800] env[62552]: DEBUG nova.compute.manager [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 775.910218] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e2ef29-de27-455a-b5cd-8a29a1920dd6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.922375] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 775.922640] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 775.922741] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Deleting the datastore file [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 775.923215] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cadb0b49-c60a-4e34-9bfc-510a6bb15d2b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.928953] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 775.928953] env[62552]: value = "task-1239282" [ 775.928953] env[62552]: _type = "Task" [ 775.928953] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.938995] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.990797] env[62552]: DEBUG nova.network.neutron [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Successfully updated port: 5cb2de18-f6f3-4351-a156-891bb4a6c542 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.030800] env[62552]: DEBUG nova.compute.manager [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 776.031392] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.031614] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.031769] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.031952] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.032108] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.032256] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.032459] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.032610] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.032770] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.032925] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.036025] env[62552]: DEBUG nova.virt.hardware [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.036232] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6ce029-a9c9-48c6-87c2-e1d576399c31 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.048164] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011e2116-131e-4cfa-a619-9229bf69ad3c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.053191] env[62552]: DEBUG nova.compute.manager [req-e105ebb2-579f-4184-a4d3-bf21427ef904 req-5b5a9ec3-25e3-45f9-8dc3-9185e46b0e5e service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Received event network-vif-plugged-5cb2de18-f6f3-4351-a156-891bb4a6c542 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 776.053442] env[62552]: DEBUG oslo_concurrency.lockutils [req-e105ebb2-579f-4184-a4d3-bf21427ef904 req-5b5a9ec3-25e3-45f9-8dc3-9185e46b0e5e service nova] Acquiring lock "ce5d0165-65f1-4505-9c46-1129c56a8913-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.053687] env[62552]: DEBUG oslo_concurrency.lockutils [req-e105ebb2-579f-4184-a4d3-bf21427ef904 req-5b5a9ec3-25e3-45f9-8dc3-9185e46b0e5e service nova] Lock "ce5d0165-65f1-4505-9c46-1129c56a8913-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.053885] env[62552]: DEBUG oslo_concurrency.lockutils [req-e105ebb2-579f-4184-a4d3-bf21427ef904 req-5b5a9ec3-25e3-45f9-8dc3-9185e46b0e5e service nova] Lock "ce5d0165-65f1-4505-9c46-1129c56a8913-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.054138] env[62552]: DEBUG nova.compute.manager [req-e105ebb2-579f-4184-a4d3-bf21427ef904 req-5b5a9ec3-25e3-45f9-8dc3-9185e46b0e5e service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] No waiting events found dispatching network-vif-plugged-5cb2de18-f6f3-4351-a156-891bb4a6c542 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 776.054634] env[62552]: WARNING nova.compute.manager [req-e105ebb2-579f-4184-a4d3-bf21427ef904 req-5b5a9ec3-25e3-45f9-8dc3-9185e46b0e5e service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Received unexpected event network-vif-plugged-5cb2de18-f6f3-4351-a156-891bb4a6c542 for instance with vm_state building and task_state spawning. [ 776.205790] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bc5cc5-f6b8-4789-b341-bbe292430cea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.213882] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98dfc6d4-28b4-409d-8879-cbdcb90294b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.248546] env[62552]: DEBUG nova.network.neutron [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Successfully created port: abe02bcf-41b7-4884-8ef3-70b5268480f0 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.251010] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427b5b26-4021-429a-a63d-cc15342f6fb2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.262984] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10482112-0ba8-40ef-a2b7-a128594c3ed9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.276908] env[62552]: DEBUG nova.compute.provider_tree [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.378049] env[62552]: DEBUG oslo_concurrency.lockutils [req-e0068a50-2c32-4cc6-a8e0-61f46f078948 req-7a537038-10e6-4330-8dc7-02b3f8e05f17 service nova] Releasing lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.439231] env[62552]: INFO nova.compute.manager [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Took 26.92 seconds to build instance. [ 776.450462] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239282, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.347932} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.450462] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 776.450462] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 776.450462] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.496076] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Acquiring lock "refresh_cache-ce5d0165-65f1-4505-9c46-1129c56a8913" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.496076] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Acquired lock "refresh_cache-ce5d0165-65f1-4505-9c46-1129c56a8913" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.496076] env[62552]: DEBUG nova.network.neutron [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.761314] env[62552]: DEBUG nova.compute.manager [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 776.781041] env[62552]: DEBUG nova.scheduler.client.report [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 776.802950] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.803222] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.803468] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.803670] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.803760] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.803909] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.804160] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.804315] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.804489] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.804691] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.804827] env[62552]: DEBUG nova.virt.hardware [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.806042] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55975dd0-f067-4770-a268-fcab567d101a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.814402] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce0434c-e450-4091-9d3c-075393407975 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.942594] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ed10443-0d02-48cd-81fd-17968686c3b7 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "8707ce4b-677e-4f13-86f8-3e327d19380b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.312s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.039723] env[62552]: DEBUG nova.network.neutron [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.235305] env[62552]: DEBUG nova.network.neutron [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Updating instance_info_cache with network_info: [{"id": "5cb2de18-f6f3-4351-a156-891bb4a6c542", "address": "fa:16:3e:6a:77:08", "network": {"id": "6b678d9d-265f-4541-a994-c70b2ba1774b", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-221581618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2fe0e0c574e4ff2a2ed4acdd4071e42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cb2de18-f6", "ovs_interfaceid": "5cb2de18-f6f3-4351-a156-891bb4a6c542", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.287318] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.560s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.287842] env[62552]: DEBUG nova.compute.manager [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 777.290332] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.769s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.445864] env[62552]: DEBUG nova.compute.manager [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 777.501251] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 777.501251] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 777.501251] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.501251] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 777.501459] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.505135] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 777.505415] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 777.505586] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 777.505756] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 777.505916] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 777.506110] env[62552]: DEBUG nova.virt.hardware [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.507305] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ad409d-b271-4704-a305-32c5d4838698 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.519783] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bed5e9-b854-48ed-b617-14a863ee59fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.537271] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.540422] env[62552]: DEBUG oslo.service.loopingcall [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.540698] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 777.540912] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-564dcd2c-709e-4e83-988a-2b8c36722de8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.559021] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.559021] env[62552]: value = "task-1239283" [ 777.559021] env[62552]: _type = "Task" [ 777.559021] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.567298] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239283, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.737698] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Releasing lock "refresh_cache-ce5d0165-65f1-4505-9c46-1129c56a8913" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.738180] env[62552]: DEBUG nova.compute.manager [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Instance network_info: |[{"id": "5cb2de18-f6f3-4351-a156-891bb4a6c542", "address": "fa:16:3e:6a:77:08", "network": {"id": "6b678d9d-265f-4541-a994-c70b2ba1774b", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-221581618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2fe0e0c574e4ff2a2ed4acdd4071e42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cb2de18-f6", "ovs_interfaceid": "5cb2de18-f6f3-4351-a156-891bb4a6c542", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 777.738666] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:77:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca50cd14-9e1f-4d74-a066-e5a45ba0ce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5cb2de18-f6f3-4351-a156-891bb4a6c542', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.747538] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Creating folder: Project (b2fe0e0c574e4ff2a2ed4acdd4071e42). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.749134] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e4f7b2f-2c86-4c1b-8dc2-5ef5abbedad8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.752780] env[62552]: DEBUG nova.compute.manager [req-93651eb3-e7ac-4616-8772-5093ec0a004a req-c93c765d-a852-44c1-8b21-f44f43e28c9c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Received event network-changed-65c6abb5-29b3-4603-b019-4716db17392c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 777.752980] env[62552]: DEBUG nova.compute.manager [req-93651eb3-e7ac-4616-8772-5093ec0a004a req-c93c765d-a852-44c1-8b21-f44f43e28c9c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Refreshing instance network info cache due to event network-changed-65c6abb5-29b3-4603-b019-4716db17392c. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 777.753208] env[62552]: DEBUG oslo_concurrency.lockutils [req-93651eb3-e7ac-4616-8772-5093ec0a004a req-c93c765d-a852-44c1-8b21-f44f43e28c9c service nova] Acquiring lock "refresh_cache-8707ce4b-677e-4f13-86f8-3e327d19380b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.753456] env[62552]: DEBUG oslo_concurrency.lockutils [req-93651eb3-e7ac-4616-8772-5093ec0a004a req-c93c765d-a852-44c1-8b21-f44f43e28c9c service nova] Acquired lock "refresh_cache-8707ce4b-677e-4f13-86f8-3e327d19380b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.753536] env[62552]: DEBUG nova.network.neutron [req-93651eb3-e7ac-4616-8772-5093ec0a004a req-c93c765d-a852-44c1-8b21-f44f43e28c9c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Refreshing network info cache for port 65c6abb5-29b3-4603-b019-4716db17392c {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 777.767240] env[62552]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 777.767432] env[62552]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62552) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 777.767819] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Folder already exists: Project (b2fe0e0c574e4ff2a2ed4acdd4071e42). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 777.768171] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Creating folder: Instances. Parent ref: group-v267354. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.768704] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-add6304e-acec-4123-ac3d-e1d9afc9a5aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.784834] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Created folder: Instances in parent group-v267354. [ 777.785446] env[62552]: DEBUG oslo.service.loopingcall [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.785847] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 777.786201] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55e4b337-4568-46cb-aba8-069722392d20 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.805862] env[62552]: DEBUG nova.compute.utils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.814361] env[62552]: DEBUG nova.compute.manager [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 777.814361] env[62552]: DEBUG nova.network.neutron [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 777.820339] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.820339] env[62552]: value = "task-1239286" [ 777.820339] env[62552]: _type = "Task" [ 777.820339] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.831199] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239286, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.939580] env[62552]: DEBUG nova.policy [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3f33c27aa38462ead7638b7cee05d45', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19ff3ed083bc42bf82311bddcd32d093', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 777.972450] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.076175] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239283, 'name': CreateVM_Task, 'duration_secs': 0.262774} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.076175] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 778.076722] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.077020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.077432] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 778.077788] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd0ff50e-54d3-43e4-accb-e1a6113256b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.082381] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 778.082381] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52749b61-aacd-5fed-e652-f54a4f924b7a" [ 778.082381] env[62552]: _type = "Task" [ 778.082381] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.093692] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52749b61-aacd-5fed-e652-f54a4f924b7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.163149] env[62552]: DEBUG nova.compute.manager [req-61978157-0368-4311-94a2-7d9adb44983d req-dcfa46d9-def5-49ba-a1a0-0041a15bc238 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Received event network-vif-plugged-abe02bcf-41b7-4884-8ef3-70b5268480f0 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 778.165132] env[62552]: DEBUG oslo_concurrency.lockutils [req-61978157-0368-4311-94a2-7d9adb44983d req-dcfa46d9-def5-49ba-a1a0-0041a15bc238 service nova] Acquiring lock "0418260a-aa27-4955-ab15-b180ec04f0b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.165132] env[62552]: DEBUG oslo_concurrency.lockutils [req-61978157-0368-4311-94a2-7d9adb44983d req-dcfa46d9-def5-49ba-a1a0-0041a15bc238 service nova] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.165132] env[62552]: DEBUG oslo_concurrency.lockutils [req-61978157-0368-4311-94a2-7d9adb44983d req-dcfa46d9-def5-49ba-a1a0-0041a15bc238 service nova] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.165132] env[62552]: DEBUG nova.compute.manager [req-61978157-0368-4311-94a2-7d9adb44983d req-dcfa46d9-def5-49ba-a1a0-0041a15bc238 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] No waiting events found dispatching network-vif-plugged-abe02bcf-41b7-4884-8ef3-70b5268480f0 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 778.165132] env[62552]: WARNING nova.compute.manager [req-61978157-0368-4311-94a2-7d9adb44983d req-dcfa46d9-def5-49ba-a1a0-0041a15bc238 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Received unexpected event network-vif-plugged-abe02bcf-41b7-4884-8ef3-70b5268480f0 for instance with vm_state building and task_state spawning. [ 778.233989] env[62552]: DEBUG nova.compute.manager [req-f5329ba5-f513-4db7-8d2f-5a0618c70acc req-f2fb7d2c-e5ce-4b9d-9067-89aaed36db27 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Received event network-changed-5cb2de18-f6f3-4351-a156-891bb4a6c542 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 778.233989] env[62552]: DEBUG nova.compute.manager [req-f5329ba5-f513-4db7-8d2f-5a0618c70acc req-f2fb7d2c-e5ce-4b9d-9067-89aaed36db27 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Refreshing instance network info cache due to event network-changed-5cb2de18-f6f3-4351-a156-891bb4a6c542. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 778.234119] env[62552]: DEBUG oslo_concurrency.lockutils [req-f5329ba5-f513-4db7-8d2f-5a0618c70acc req-f2fb7d2c-e5ce-4b9d-9067-89aaed36db27 service nova] Acquiring lock "refresh_cache-ce5d0165-65f1-4505-9c46-1129c56a8913" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.234187] env[62552]: DEBUG oslo_concurrency.lockutils [req-f5329ba5-f513-4db7-8d2f-5a0618c70acc req-f2fb7d2c-e5ce-4b9d-9067-89aaed36db27 service nova] Acquired lock "refresh_cache-ce5d0165-65f1-4505-9c46-1129c56a8913" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.234408] env[62552]: DEBUG nova.network.neutron [req-f5329ba5-f513-4db7-8d2f-5a0618c70acc req-f2fb7d2c-e5ce-4b9d-9067-89aaed36db27 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Refreshing network info cache for port 5cb2de18-f6f3-4351-a156-891bb4a6c542 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 778.259875] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2bded0-68cb-4fba-863a-196ddf3105e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.268172] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2708ad28-a72c-4692-83fa-cb37122c228f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.300984] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000955b3-9704-4765-bbda-f4e72607f1ec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.308716] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82fa972f-5663-450d-9e1a-8d554da45b36 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.314578] env[62552]: DEBUG nova.compute.manager [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 778.325216] env[62552]: DEBUG nova.compute.provider_tree [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.337431] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239286, 'name': CreateVM_Task, 'duration_secs': 0.32962} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.337580] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 778.338283] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267357', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'name': 'volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce5d0165-65f1-4505-9c46-1129c56a8913', 'attached_at': '', 'detached_at': '', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'serial': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed'}, 'disk_bus': None, 'boot_index': 0, 'attachment_id': '92799767-511b-4d29-af4c-507a783b74dc', 'volume_type': None}], 'swap': None} {{(pid=62552) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 778.338478] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Root volume attach. Driver type: vmdk {{(pid=62552) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 778.339278] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4be574b-c244-45d0-961e-905f4f7e6ee4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.347086] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22f5b6a-8856-44f4-89d3-aaec0963e09a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.353183] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7c3d7d-ec62-40c6-8338-4c73551876b8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.359758] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-96901963-cfe7-4104-a80a-e8d18a07a0fb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.367076] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for the task: (returnval){ [ 778.367076] env[62552]: value = "task-1239287" [ 778.367076] env[62552]: _type = "Task" [ 778.367076] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.377036] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239287, 'name': RelocateVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.538214] env[62552]: DEBUG nova.network.neutron [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Successfully created port: 7378d663-07fd-4412-9093-8774fdf7c753 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.596301] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52749b61-aacd-5fed-e652-f54a4f924b7a, 'name': SearchDatastore_Task, 'duration_secs': 0.011404} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.596645] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.596966] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.597282] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.597441] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.597641] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.597901] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2ea0b63-76b7-42f6-a780-5058795245a5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.605718] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.605888] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 778.606756] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9cbf7881-5898-4b79-b933-ee4283ef7a91 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.611997] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 778.611997] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ab8c42-f214-df55-6120-0868529dcca4" [ 778.611997] env[62552]: _type = "Task" [ 778.611997] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.620039] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ab8c42-f214-df55-6120-0868529dcca4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.625610] env[62552]: DEBUG nova.network.neutron [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Successfully updated port: abe02bcf-41b7-4884-8ef3-70b5268480f0 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 778.832870] env[62552]: DEBUG nova.scheduler.client.report [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 778.883674] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239287, 'name': RelocateVM_Task} progress is 20%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.122791] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ab8c42-f214-df55-6120-0868529dcca4, 'name': SearchDatastore_Task, 'duration_secs': 0.008118} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.123611] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed1885ae-0aa1-4889-8930-14cc8b8253dd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.127530] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.127666] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.127802] env[62552]: DEBUG nova.network.neutron [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.130211] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 779.130211] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52849431-ebc2-a497-8280-93086dde9d91" [ 779.130211] env[62552]: _type = "Task" [ 779.130211] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.142058] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52849431-ebc2-a497-8280-93086dde9d91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.156684] env[62552]: DEBUG nova.network.neutron [req-93651eb3-e7ac-4616-8772-5093ec0a004a req-c93c765d-a852-44c1-8b21-f44f43e28c9c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Updated VIF entry in instance network info cache for port 65c6abb5-29b3-4603-b019-4716db17392c. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 779.157068] env[62552]: DEBUG nova.network.neutron [req-93651eb3-e7ac-4616-8772-5093ec0a004a req-c93c765d-a852-44c1-8b21-f44f43e28c9c service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Updating instance_info_cache with network_info: [{"id": "65c6abb5-29b3-4603-b019-4716db17392c", "address": "fa:16:3e:2d:a8:73", "network": {"id": "7e24b87f-25ff-4556-a4df-5cf01e529ac5", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-395831708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "56e85e5029e347e89447147bfc0ee79b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c6abb5-29", "ovs_interfaceid": "65c6abb5-29b3-4603-b019-4716db17392c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.338412] env[62552]: DEBUG nova.compute.manager [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 779.341514] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.051s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.342234] env[62552]: ERROR nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896, please check neutron logs for more information. [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Traceback (most recent call last): [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self.driver.spawn(context, instance, image_meta, [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] vm_ref = self.build_virtual_machine(instance, [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.342234] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] for vif in network_info: [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] return self._sync_wrapper(fn, *args, **kwargs) [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self.wait() [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self[:] = self._gt.wait() [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] return self._exit_event.wait() [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] result = hub.switch() [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 779.342625] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] return self.greenlet.switch() [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] result = function(*args, **kwargs) [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] return func(*args, **kwargs) [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] raise e [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] nwinfo = self.network_api.allocate_for_instance( [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] created_port_ids = self._update_ports_for_instance( [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] with excutils.save_and_reraise_exception(): [ 779.343030] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] self.force_reraise() [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] raise self.value [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] updated_port = self._update_port( [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] _ensure_no_port_binding_failure(port) [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] raise exception.PortBindingFailed(port_id=port['id']) [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] nova.exception.PortBindingFailed: Binding failed for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896, please check neutron logs for more information. [ 779.343459] env[62552]: ERROR nova.compute.manager [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] [ 779.343794] env[62552]: DEBUG nova.compute.utils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Binding failed for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.346357] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.822s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.352832] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Build of instance ab363b52-ade2-45c4-ba3a-ae46ffaf831e was re-scheduled: Binding failed for port 4ef84cd8-05a5-4f97-8e3e-020d99a73896, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 779.355318] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 779.355318] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Acquiring lock "refresh_cache-ab363b52-ade2-45c4-ba3a-ae46ffaf831e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.355318] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Acquired lock "refresh_cache-ab363b52-ade2-45c4-ba3a-ae46ffaf831e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.355318] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.378472] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.378872] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.379223] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.379223] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.379465] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.379668] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.379949] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.380187] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.380634] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.380634] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.380831] env[62552]: DEBUG nova.virt.hardware [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.381950] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18717072-5c07-407d-ad7e-8fb97f790370 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.388786] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239287, 'name': RelocateVM_Task} progress is 20%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.395011] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd75111e-65b7-4103-a199-6251c859b7b5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.523148] env[62552]: DEBUG nova.network.neutron [req-f5329ba5-f513-4db7-8d2f-5a0618c70acc req-f2fb7d2c-e5ce-4b9d-9067-89aaed36db27 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Updated VIF entry in instance network info cache for port 5cb2de18-f6f3-4351-a156-891bb4a6c542. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 779.523528] env[62552]: DEBUG nova.network.neutron [req-f5329ba5-f513-4db7-8d2f-5a0618c70acc req-f2fb7d2c-e5ce-4b9d-9067-89aaed36db27 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Updating instance_info_cache with network_info: [{"id": "5cb2de18-f6f3-4351-a156-891bb4a6c542", "address": "fa:16:3e:6a:77:08", "network": {"id": "6b678d9d-265f-4541-a994-c70b2ba1774b", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-221581618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2fe0e0c574e4ff2a2ed4acdd4071e42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cb2de18-f6", "ovs_interfaceid": "5cb2de18-f6f3-4351-a156-891bb4a6c542", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.643515] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52849431-ebc2-a497-8280-93086dde9d91, 'name': SearchDatastore_Task, 'duration_secs': 0.032699} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.643867] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.644160] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282/1472ca01-ba04-4e7d-a9cf-ccaf29c04282.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 779.644442] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f408b26c-5d9a-464f-8b9e-8e205ba8703e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.651637] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 779.651637] env[62552]: value = "task-1239288" [ 779.651637] env[62552]: _type = "Task" [ 779.651637] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.660308] env[62552]: DEBUG oslo_concurrency.lockutils [req-93651eb3-e7ac-4616-8772-5093ec0a004a req-c93c765d-a852-44c1-8b21-f44f43e28c9c service nova] Releasing lock "refresh_cache-8707ce4b-677e-4f13-86f8-3e327d19380b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.660776] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239288, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.689113] env[62552]: DEBUG nova.network.neutron [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.882975] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239287, 'name': RelocateVM_Task} progress is 20%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.891666] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.950706] env[62552]: DEBUG nova.network.neutron [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updating instance_info_cache with network_info: [{"id": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "address": "fa:16:3e:82:10:f7", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabe02bcf-41", "ovs_interfaceid": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.026767] env[62552]: DEBUG oslo_concurrency.lockutils [req-f5329ba5-f513-4db7-8d2f-5a0618c70acc req-f2fb7d2c-e5ce-4b9d-9067-89aaed36db27 service nova] Releasing lock "refresh_cache-ce5d0165-65f1-4505-9c46-1129c56a8913" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.038073] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.162580] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239288, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.241195] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3c93e7-a61e-430c-b820-06b561c683e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.249394] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd30d24f-faf7-453f-a7ab-fae4752c219c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.255552] env[62552]: DEBUG nova.compute.manager [req-de436f2a-bc53-4d7d-a4ef-9ead704d3c67 req-b4f69f6b-6b44-4f07-a4c8-67a4a5cb3391 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Received event network-changed-abe02bcf-41b7-4884-8ef3-70b5268480f0 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 780.256272] env[62552]: DEBUG nova.compute.manager [req-de436f2a-bc53-4d7d-a4ef-9ead704d3c67 req-b4f69f6b-6b44-4f07-a4c8-67a4a5cb3391 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Refreshing instance network info cache due to event network-changed-abe02bcf-41b7-4884-8ef3-70b5268480f0. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 780.256272] env[62552]: DEBUG oslo_concurrency.lockutils [req-de436f2a-bc53-4d7d-a4ef-9ead704d3c67 req-b4f69f6b-6b44-4f07-a4c8-67a4a5cb3391 service nova] Acquiring lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.282583] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff554ac-51cc-4312-94ac-ff47f04204ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.289767] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811c5780-0c5d-4f5e-be80-6c9ade4132bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.305944] env[62552]: DEBUG nova.compute.provider_tree [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.383846] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239287, 'name': RelocateVM_Task} progress is 20%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.455885] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.456291] env[62552]: DEBUG nova.compute.manager [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Instance network_info: |[{"id": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "address": "fa:16:3e:82:10:f7", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabe02bcf-41", "ovs_interfaceid": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 780.456599] env[62552]: DEBUG oslo_concurrency.lockutils [req-de436f2a-bc53-4d7d-a4ef-9ead704d3c67 req-b4f69f6b-6b44-4f07-a4c8-67a4a5cb3391 service nova] Acquired lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.456778] env[62552]: DEBUG nova.network.neutron [req-de436f2a-bc53-4d7d-a4ef-9ead704d3c67 req-b4f69f6b-6b44-4f07-a4c8-67a4a5cb3391 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Refreshing network info cache for port abe02bcf-41b7-4884-8ef3-70b5268480f0 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 780.458014] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:10:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'abe02bcf-41b7-4884-8ef3-70b5268480f0', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 780.465804] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Creating folder: Project (3758acde6e3e4320bf63d4cd1667bfc2). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 780.466838] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2886a7ff-123e-4f1f-a6ad-96e3deb3adaa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.477722] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Created folder: Project (3758acde6e3e4320bf63d4cd1667bfc2) in parent group-v267339. [ 780.477899] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Creating folder: Instances. Parent ref: group-v267371. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 780.478152] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c1d26018-01ec-4be5-bb9a-9299a7ccb51c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.487045] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Created folder: Instances in parent group-v267371. [ 780.487045] env[62552]: DEBUG oslo.service.loopingcall [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.487045] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 780.487233] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-683d8133-5cdd-4631-b629-1e8b7305daa7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.506021] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.506021] env[62552]: value = "task-1239291" [ 780.506021] env[62552]: _type = "Task" [ 780.506021] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.513547] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239291, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.540967] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Releasing lock "refresh_cache-ab363b52-ade2-45c4-ba3a-ae46ffaf831e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.541277] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 780.541635] env[62552]: DEBUG nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 780.542669] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.566518] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.662715] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239288, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.614633} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.663052] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282/1472ca01-ba04-4e7d-a9cf-ccaf29c04282.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 780.663183] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.663439] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14dac37c-2ef1-4d50-a3fa-aaff898bb2c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.669863] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 780.669863] env[62552]: value = "task-1239292" [ 780.669863] env[62552]: _type = "Task" [ 780.669863] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.677897] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239292, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.809294] env[62552]: DEBUG nova.scheduler.client.report [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 780.883960] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239287, 'name': RelocateVM_Task, 'duration_secs': 2.452845} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.884586] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 780.884955] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267357', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'name': 'volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce5d0165-65f1-4505-9c46-1129c56a8913', 'attached_at': '', 'detached_at': '', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'serial': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 780.885573] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81086052-96b2-4155-9277-a3a4cc46dd30 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.904376] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8836758-e87f-45d6-ad08-1d4cef8851a1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.926990] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed/volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.927341] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a275a22-97c1-43b9-8b0d-2fb06be16ca2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.946485] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for the task: (returnval){ [ 780.946485] env[62552]: value = "task-1239293" [ 780.946485] env[62552]: _type = "Task" [ 780.946485] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.956115] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239293, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.017790] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239291, 'name': CreateVM_Task, 'duration_secs': 0.424318} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.017887] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 781.018588] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.018743] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.019098] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 781.019347] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e72347f-1ce5-4ad8-a8a3-e625bddb3f67 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.023960] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 781.023960] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a5342d-d96d-15d8-b764-f47dc89ca6fe" [ 781.023960] env[62552]: _type = "Task" [ 781.023960] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.031828] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a5342d-d96d-15d8-b764-f47dc89ca6fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.069067] env[62552]: DEBUG nova.network.neutron [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.072053] env[62552]: DEBUG nova.network.neutron [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Successfully updated port: 7378d663-07fd-4412-9093-8774fdf7c753 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 781.181244] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239292, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076918} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.181536] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.182492] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da42756b-7714-4968-b800-01a00d111256 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.201818] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282/1472ca01-ba04-4e7d-a9cf-ccaf29c04282.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.204327] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7aafbb9a-f8bb-4c8a-9f9b-657c5ddff4eb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.223877] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 781.223877] env[62552]: value = "task-1239294" [ 781.223877] env[62552]: _type = "Task" [ 781.223877] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.231763] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239294, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.294665] env[62552]: DEBUG nova.network.neutron [req-de436f2a-bc53-4d7d-a4ef-9ead704d3c67 req-b4f69f6b-6b44-4f07-a4c8-67a4a5cb3391 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updated VIF entry in instance network info cache for port abe02bcf-41b7-4884-8ef3-70b5268480f0. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 781.295157] env[62552]: DEBUG nova.network.neutron [req-de436f2a-bc53-4d7d-a4ef-9ead704d3c67 req-b4f69f6b-6b44-4f07-a4c8-67a4a5cb3391 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updating instance_info_cache with network_info: [{"id": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "address": "fa:16:3e:82:10:f7", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabe02bcf-41", "ovs_interfaceid": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.314665] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.315282] env[62552]: ERROR nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614, please check neutron logs for more information. [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Traceback (most recent call last): [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self.driver.spawn(context, instance, image_meta, [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] vm_ref = self.build_virtual_machine(instance, [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.315282] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] for vif in network_info: [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] return self._sync_wrapper(fn, *args, **kwargs) [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self.wait() [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self[:] = self._gt.wait() [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] return self._exit_event.wait() [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] result = hub.switch() [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 781.315702] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] return self.greenlet.switch() [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] result = function(*args, **kwargs) [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] return func(*args, **kwargs) [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] raise e [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] nwinfo = self.network_api.allocate_for_instance( [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] created_port_ids = self._update_ports_for_instance( [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] with excutils.save_and_reraise_exception(): [ 781.316214] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] self.force_reraise() [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] raise self.value [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] updated_port = self._update_port( [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] _ensure_no_port_binding_failure(port) [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] raise exception.PortBindingFailed(port_id=port['id']) [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] nova.exception.PortBindingFailed: Binding failed for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614, please check neutron logs for more information. [ 781.316691] env[62552]: ERROR nova.compute.manager [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] [ 781.317084] env[62552]: DEBUG nova.compute.utils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Binding failed for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.317221] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.729s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.320368] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Build of instance 1c7e408a-392d-4fcf-bae8-588ed9f92fec was re-scheduled: Binding failed for port 5e01d6b7-47ac-44d6-bf2f-097cb2207614, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 781.320787] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 781.321011] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Acquiring lock "refresh_cache-1c7e408a-392d-4fcf-bae8-588ed9f92fec" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.321198] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Acquired lock "refresh_cache-1c7e408a-392d-4fcf-bae8-588ed9f92fec" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.321363] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.457393] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239293, 'name': ReconfigVM_Task, 'duration_secs': 0.276104} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.457710] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Reconfigured VM instance instance-00000030 to attach disk [datastore2] volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed/volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.462661] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71a202e0-fe77-4c88-84db-ce3c74d492a1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.479090] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for the task: (returnval){ [ 781.479090] env[62552]: value = "task-1239295" [ 781.479090] env[62552]: _type = "Task" [ 781.479090] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.488291] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.534091] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a5342d-d96d-15d8-b764-f47dc89ca6fe, 'name': SearchDatastore_Task, 'duration_secs': 0.008758} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.534091] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.534305] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.534353] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.534493] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.534657] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.534907] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a397c74-29ee-4757-83e1-13d904eb0028 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.542518] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.542701] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 781.543452] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a62c1cf-400c-467d-a79a-746ed9e94e70 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.548299] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 781.548299] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ebcb98-406c-670c-85e7-2f7a5de538a0" [ 781.548299] env[62552]: _type = "Task" [ 781.548299] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.555593] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ebcb98-406c-670c-85e7-2f7a5de538a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.571351] env[62552]: INFO nova.compute.manager [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] [instance: ab363b52-ade2-45c4-ba3a-ae46ffaf831e] Took 1.03 seconds to deallocate network for instance. [ 781.574897] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "refresh_cache-00b06df7-e7ae-48fd-8887-65647931d5cf" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.574991] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquired lock "refresh_cache-00b06df7-e7ae-48fd-8887-65647931d5cf" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.575241] env[62552]: DEBUG nova.network.neutron [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.735055] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239294, 'name': ReconfigVM_Task, 'duration_secs': 0.270156} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.735406] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282/1472ca01-ba04-4e7d-a9cf-ccaf29c04282.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.735737] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-504b92f2-3f44-4336-9f7d-da5d94f5f9b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.743296] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 781.743296] env[62552]: value = "task-1239296" [ 781.743296] env[62552]: _type = "Task" [ 781.743296] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.751196] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239296, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.797502] env[62552]: DEBUG oslo_concurrency.lockutils [req-de436f2a-bc53-4d7d-a4ef-9ead704d3c67 req-b4f69f6b-6b44-4f07-a4c8-67a4a5cb3391 service nova] Releasing lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.841838] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.956111] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.989570] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239295, 'name': ReconfigVM_Task, 'duration_secs': 0.273936} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.989924] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267357', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'name': 'volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce5d0165-65f1-4505-9c46-1129c56a8913', 'attached_at': '', 'detached_at': '', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'serial': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 781.990482] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2bb96ed-da65-4c0d-ad5f-30aea9f969bf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.996811] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for the task: (returnval){ [ 781.996811] env[62552]: value = "task-1239297" [ 781.996811] env[62552]: _type = "Task" [ 781.996811] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.004701] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239297, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.061641] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ebcb98-406c-670c-85e7-2f7a5de538a0, 'name': SearchDatastore_Task, 'duration_secs': 0.008956} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.062257] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30d2bf03-3560-4a34-8295-ec3dbf3a71b5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.069791] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 782.069791] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529013ae-7b24-6910-2e23-d69d4f11c342" [ 782.069791] env[62552]: _type = "Task" [ 782.069791] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.081939] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529013ae-7b24-6910-2e23-d69d4f11c342, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.115924] env[62552]: DEBUG nova.network.neutron [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.170500] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73053b68-a20f-4293-b0d3-2f2974aa324f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.178147] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfa7b59-95f8-4819-ab82-ae9b1e42e8cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.209608] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a939ebc6-34cd-4206-bb0d-3d9750a523fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.216899] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59008662-96d9-4132-b70e-cd9e46b88e07 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.231670] env[62552]: DEBUG nova.compute.provider_tree [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.252504] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239296, 'name': Rename_Task, 'duration_secs': 0.136978} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.255118] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.255382] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73985a01-1f0a-4313-bef2-c9aa41179226 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.262880] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 782.262880] env[62552]: value = "task-1239298" [ 782.262880] env[62552]: _type = "Task" [ 782.262880] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.270795] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239298, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.312311] env[62552]: DEBUG nova.compute.manager [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Received event network-vif-plugged-7378d663-07fd-4412-9093-8774fdf7c753 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 782.312311] env[62552]: DEBUG oslo_concurrency.lockutils [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] Acquiring lock "00b06df7-e7ae-48fd-8887-65647931d5cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.312311] env[62552]: DEBUG oslo_concurrency.lockutils [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.312700] env[62552]: DEBUG oslo_concurrency.lockutils [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.312700] env[62552]: DEBUG nova.compute.manager [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] No waiting events found dispatching network-vif-plugged-7378d663-07fd-4412-9093-8774fdf7c753 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 782.312700] env[62552]: WARNING nova.compute.manager [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Received unexpected event network-vif-plugged-7378d663-07fd-4412-9093-8774fdf7c753 for instance with vm_state building and task_state spawning. [ 782.312858] env[62552]: DEBUG nova.compute.manager [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Received event network-changed-7378d663-07fd-4412-9093-8774fdf7c753 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 782.312982] env[62552]: DEBUG nova.compute.manager [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Refreshing instance network info cache due to event network-changed-7378d663-07fd-4412-9093-8774fdf7c753. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 782.313166] env[62552]: DEBUG oslo_concurrency.lockutils [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] Acquiring lock "refresh_cache-00b06df7-e7ae-48fd-8887-65647931d5cf" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.362340] env[62552]: DEBUG nova.network.neutron [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Updating instance_info_cache with network_info: [{"id": "7378d663-07fd-4412-9093-8774fdf7c753", "address": "fa:16:3e:4c:d7:af", "network": {"id": "2bf558f3-1ebd-4738-8e1f-10081487cda2", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1268991274-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ff3ed083bc42bf82311bddcd32d093", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7378d663-07", "ovs_interfaceid": "7378d663-07fd-4412-9093-8774fdf7c753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.459142] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Releasing lock "refresh_cache-1c7e408a-392d-4fcf-bae8-588ed9f92fec" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.459399] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 782.459582] env[62552]: DEBUG nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 782.459750] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.475011] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.506924] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239297, 'name': Rename_Task, 'duration_secs': 0.158668} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.507260] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.507519] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5063958e-756d-4071-852f-0a4a4b63360a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.514541] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for the task: (returnval){ [ 782.514541] env[62552]: value = "task-1239299" [ 782.514541] env[62552]: _type = "Task" [ 782.514541] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.524676] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239299, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.580410] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529013ae-7b24-6910-2e23-d69d4f11c342, 'name': SearchDatastore_Task, 'duration_secs': 0.00814} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.580695] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.580955] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 0418260a-aa27-4955-ab15-b180ec04f0b3/0418260a-aa27-4955-ab15-b180ec04f0b3.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 782.581240] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a1b349d-8eb8-423e-a9be-8800e3d8122c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.589699] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 782.589699] env[62552]: value = "task-1239300" [ 782.589699] env[62552]: _type = "Task" [ 782.589699] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.597666] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239300, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.604637] env[62552]: INFO nova.scheduler.client.report [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Deleted allocations for instance ab363b52-ade2-45c4-ba3a-ae46ffaf831e [ 782.736016] env[62552]: DEBUG nova.scheduler.client.report [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 782.773624] env[62552]: DEBUG oslo_vmware.api [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239298, 'name': PowerOnVM_Task, 'duration_secs': 0.410355} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.773907] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 782.774217] env[62552]: DEBUG nova.compute.manager [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 782.775165] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edeb2d0-c9f9-404c-b70b-b98579468fb7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.864905] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Releasing lock "refresh_cache-00b06df7-e7ae-48fd-8887-65647931d5cf" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.864905] env[62552]: DEBUG nova.compute.manager [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Instance network_info: |[{"id": "7378d663-07fd-4412-9093-8774fdf7c753", "address": "fa:16:3e:4c:d7:af", "network": {"id": "2bf558f3-1ebd-4738-8e1f-10081487cda2", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1268991274-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ff3ed083bc42bf82311bddcd32d093", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7378d663-07", "ovs_interfaceid": "7378d663-07fd-4412-9093-8774fdf7c753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 782.865191] env[62552]: DEBUG oslo_concurrency.lockutils [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] Acquired lock "refresh_cache-00b06df7-e7ae-48fd-8887-65647931d5cf" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.865254] env[62552]: DEBUG nova.network.neutron [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Refreshing network info cache for port 7378d663-07fd-4412-9093-8774fdf7c753 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 782.866510] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:d7:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7378d663-07fd-4412-9093-8774fdf7c753', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.876027] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Creating folder: Project (19ff3ed083bc42bf82311bddcd32d093). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.877150] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a9d36c9-a752-4117-8e27-ab16616c4823 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.888181] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Created folder: Project (19ff3ed083bc42bf82311bddcd32d093) in parent group-v267339. [ 782.888700] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Creating folder: Instances. Parent ref: group-v267374. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.888700] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9869b11-55c4-4cbf-bc3d-6e45c25a1154 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.900016] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Created folder: Instances in parent group-v267374. [ 782.900016] env[62552]: DEBUG oslo.service.loopingcall [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.900194] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.900434] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77eaca2a-dae0-45d9-b541-6068c2fe5cf3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.919648] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.919648] env[62552]: value = "task-1239303" [ 782.919648] env[62552]: _type = "Task" [ 782.919648] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.927605] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239303, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.977987] env[62552]: DEBUG nova.network.neutron [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.024656] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239299, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.099488] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239300, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432371} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.099488] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 0418260a-aa27-4955-ab15-b180ec04f0b3/0418260a-aa27-4955-ab15-b180ec04f0b3.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 783.099844] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.099844] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a683219-e0ef-4970-a1d2-ce585d2fe18c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.106174] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 783.106174] env[62552]: value = "task-1239304" [ 783.106174] env[62552]: _type = "Task" [ 783.106174] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.114006] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239304, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.115597] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6daf9c1-a084-480b-b54f-1db013a67671 tempest-ServersNegativeTestMultiTenantJSON-2122499842 tempest-ServersNegativeTestMultiTenantJSON-2122499842-project-member] Lock "ab363b52-ade2-45c4-ba3a-ae46ffaf831e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 178.647s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.244588] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.244588] env[62552]: ERROR nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4eb4b194-659d-40f4-b9fb-f110b64965fa, please check neutron logs for more information. [ 783.244588] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Traceback (most recent call last): [ 783.244588] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 783.244588] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self.driver.spawn(context, instance, image_meta, [ 783.244588] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 783.244588] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.244588] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.244588] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] vm_ref = self.build_virtual_machine(instance, [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] for vif in network_info: [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] return self._sync_wrapper(fn, *args, **kwargs) [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self.wait() [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self[:] = self._gt.wait() [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] return self._exit_event.wait() [ 783.245374] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] result = hub.switch() [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] return self.greenlet.switch() [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] result = function(*args, **kwargs) [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] return func(*args, **kwargs) [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] raise e [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] nwinfo = self.network_api.allocate_for_instance( [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.245742] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] created_port_ids = self._update_ports_for_instance( [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] with excutils.save_and_reraise_exception(): [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] self.force_reraise() [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] raise self.value [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] updated_port = self._update_port( [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] _ensure_no_port_binding_failure(port) [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.246177] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] raise exception.PortBindingFailed(port_id=port['id']) [ 783.246609] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] nova.exception.PortBindingFailed: Binding failed for port 4eb4b194-659d-40f4-b9fb-f110b64965fa, please check neutron logs for more information. [ 783.246609] env[62552]: ERROR nova.compute.manager [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] [ 783.246609] env[62552]: DEBUG nova.compute.utils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Binding failed for port 4eb4b194-659d-40f4-b9fb-f110b64965fa, please check neutron logs for more information. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 783.246609] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.366s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.246609] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.246782] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 783.247149] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.876s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.248906] env[62552]: INFO nova.compute.claims [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.252112] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Build of instance 1252f45a-d55d-44c0-8629-aba51fb89956 was re-scheduled: Binding failed for port 4eb4b194-659d-40f4-b9fb-f110b64965fa, please check neutron logs for more information. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 783.252456] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Unplugging VIFs for instance {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 783.252743] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Acquiring lock "refresh_cache-1252f45a-d55d-44c0-8629-aba51fb89956" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.252806] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Acquired lock "refresh_cache-1252f45a-d55d-44c0-8629-aba51fb89956" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.252963] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.255039] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6910cfc4-86fa-4115-a1a4-7d9e0f574cfc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.263500] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2adbe5e-e2c8-47b1-b73c-f36ee709e82f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.277810] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-495bce70-fb5c-4c65-8340-466fad517f8d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.284349] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987a7845-4d89-4aca-bacc-8f265c6eadb0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.295540] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.319233] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181523MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 783.319374] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.429864] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239303, 'name': CreateVM_Task, 'duration_secs': 0.299173} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.430275] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 783.430977] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.431161] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.431498] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 783.431749] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b229617-64cb-4580-977a-bba3de6928f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.438458] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 783.438458] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52add642-aae1-9b38-c60e-76784a2ddad8" [ 783.438458] env[62552]: _type = "Task" [ 783.438458] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.449993] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52add642-aae1-9b38-c60e-76784a2ddad8, 'name': SearchDatastore_Task, 'duration_secs': 0.008461} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.452390] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.452971] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.452971] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.452971] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.453552] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.453552] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-60559fe3-3925-43b3-a1c8-78c3aef22310 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.464836] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.465018] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.465828] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c729fb83-5b7c-44bf-90e8-96d5db5ef004 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.471363] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 783.471363] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5227041b-af2c-fb7a-0f51-d171572de696" [ 783.471363] env[62552]: _type = "Task" [ 783.471363] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.479558] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5227041b-af2c-fb7a-0f51-d171572de696, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.482227] env[62552]: INFO nova.compute.manager [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] [instance: 1c7e408a-392d-4fcf-bae8-588ed9f92fec] Took 1.02 seconds to deallocate network for instance. [ 783.525184] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239299, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.619023] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239304, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063255} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.619023] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.619023] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c501b5a-79e3-4ba0-b678-3e279980cc40 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.619444] env[62552]: DEBUG nova.compute.manager [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 783.644083] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 0418260a-aa27-4955-ab15-b180ec04f0b3/0418260a-aa27-4955-ab15-b180ec04f0b3.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.644859] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18f42fe7-0f16-4be9-b20b-6e3aafc2d263 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.668024] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 783.668024] env[62552]: value = "task-1239305" [ 783.668024] env[62552]: _type = "Task" [ 783.668024] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.675706] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239305, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.802345] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.983725] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5227041b-af2c-fb7a-0f51-d171572de696, 'name': SearchDatastore_Task, 'duration_secs': 0.046171} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.990018] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67e04758-c906-49cb-9e6e-e1a72aa34db3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.993984] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 783.993984] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1d33f-a229-7cdf-869a-00b69ace0a3d" [ 783.993984] env[62552]: _type = "Task" [ 783.993984] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.002681] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "1472ca01-ba04-4e7d-a9cf-ccaf29c04282" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.003165] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "1472ca01-ba04-4e7d-a9cf-ccaf29c04282" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.003495] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "1472ca01-ba04-4e7d-a9cf-ccaf29c04282-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.003816] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "1472ca01-ba04-4e7d-a9cf-ccaf29c04282-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.004137] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "1472ca01-ba04-4e7d-a9cf-ccaf29c04282-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.006561] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1d33f-a229-7cdf-869a-00b69ace0a3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.006561] env[62552]: INFO nova.compute.manager [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Terminating instance [ 784.048302] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239299, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.075270] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.141150] env[62552]: DEBUG nova.network.neutron [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Updated VIF entry in instance network info cache for port 7378d663-07fd-4412-9093-8774fdf7c753. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 784.141542] env[62552]: DEBUG nova.network.neutron [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Updating instance_info_cache with network_info: [{"id": "7378d663-07fd-4412-9093-8774fdf7c753", "address": "fa:16:3e:4c:d7:af", "network": {"id": "2bf558f3-1ebd-4738-8e1f-10081487cda2", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1268991274-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ff3ed083bc42bf82311bddcd32d093", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7378d663-07", "ovs_interfaceid": "7378d663-07fd-4412-9093-8774fdf7c753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.152121] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.176655] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.507048] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1d33f-a229-7cdf-869a-00b69ace0a3d, 'name': SearchDatastore_Task, 'duration_secs': 0.021234} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.507355] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.507890] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 00b06df7-e7ae-48fd-8887-65647931d5cf/00b06df7-e7ae-48fd-8887-65647931d5cf.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.507986] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5037866c-9a50-4b2c-9b45-dad39c41d979 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.511245] env[62552]: INFO nova.scheduler.client.report [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Deleted allocations for instance 1c7e408a-392d-4fcf-bae8-588ed9f92fec [ 784.520212] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "refresh_cache-1472ca01-ba04-4e7d-a9cf-ccaf29c04282" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.520392] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquired lock "refresh_cache-1472ca01-ba04-4e7d-a9cf-ccaf29c04282" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.520583] env[62552]: DEBUG nova.network.neutron [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 784.535325] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 784.535325] env[62552]: value = "task-1239306" [ 784.535325] env[62552]: _type = "Task" [ 784.535325] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.554859] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239306, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.558643] env[62552]: DEBUG oslo_vmware.api [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239299, 'name': PowerOnVM_Task, 'duration_secs': 1.611654} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.561634] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 784.561849] env[62552]: INFO nova.compute.manager [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Took 8.53 seconds to spawn the instance on the hypervisor. [ 784.562036] env[62552]: DEBUG nova.compute.manager [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 784.563087] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1641965-990f-4e66-8e66-d541afafb993 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.578522] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Releasing lock "refresh_cache-1252f45a-d55d-44c0-8629-aba51fb89956" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.578787] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62552) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 784.578985] env[62552]: DEBUG nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 784.579159] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.604832] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.645465] env[62552]: DEBUG oslo_concurrency.lockutils [req-cbc3911d-834a-486a-b23d-8a0ea6c738dd req-5a16eb30-3b86-4766-9844-f1cfb7da1e8c service nova] Releasing lock "refresh_cache-00b06df7-e7ae-48fd-8887-65647931d5cf" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.657582] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7bbf47-3106-4e9f-b926-27f74614e02e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.664451] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d29787-c65b-4854-bfde-a70693ea10bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.674648] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239305, 'name': ReconfigVM_Task, 'duration_secs': 0.9551} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.699238] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 0418260a-aa27-4955-ab15-b180ec04f0b3/0418260a-aa27-4955-ab15-b180ec04f0b3.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.700332] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de54c4cd-8d21-433e-997b-73f4b48b30f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.702357] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3944a14b-3519-4b67-94d7-ff333c8a83af {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.710948] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-315dd36c-f82b-463b-b15b-20d035447f40 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.714422] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 784.714422] env[62552]: value = "task-1239307" [ 784.714422] env[62552]: _type = "Task" [ 784.714422] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.725955] env[62552]: DEBUG nova.compute.provider_tree [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.731597] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239307, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.032572] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f611bb3-e803-412b-b54a-2996d9a881bd tempest-InstanceActionsV221TestJSON-158309157 tempest-InstanceActionsV221TestJSON-158309157-project-member] Lock "1c7e408a-392d-4fcf-bae8-588ed9f92fec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 173.141s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.045468] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239306, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.058465] env[62552]: DEBUG nova.network.neutron [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.084407] env[62552]: INFO nova.compute.manager [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Took 31.23 seconds to build instance. [ 785.107492] env[62552]: DEBUG nova.network.neutron [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.192968] env[62552]: DEBUG nova.network.neutron [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.225471] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239307, 'name': Rename_Task, 'duration_secs': 0.151575} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.225691] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 785.225967] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8334b3f4-10d5-4a2d-a044-21f5330a53e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.228612] env[62552]: DEBUG nova.scheduler.client.report [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 785.236923] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 785.236923] env[62552]: value = "task-1239308" [ 785.236923] env[62552]: _type = "Task" [ 785.236923] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.246621] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239308, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.538480] env[62552]: DEBUG nova.compute.manager [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 785.550553] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239306, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615759} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.550828] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 00b06df7-e7ae-48fd-8887-65647931d5cf/00b06df7-e7ae-48fd-8887-65647931d5cf.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 785.551052] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.551328] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c29fd9d-0b99-4d65-8c4a-f663ed72789f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.557776] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 785.557776] env[62552]: value = "task-1239309" [ 785.557776] env[62552]: _type = "Task" [ 785.557776] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.567127] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239309, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.586633] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7702a8c-280d-4fdb-af40-cdea12479eac tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "ce5d0165-65f1-4505-9c46-1129c56a8913" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.764s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.609895] env[62552]: INFO nova.compute.manager [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] [instance: 1252f45a-d55d-44c0-8629-aba51fb89956] Took 1.03 seconds to deallocate network for instance. [ 785.696031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Releasing lock "refresh_cache-1472ca01-ba04-4e7d-a9cf-ccaf29c04282" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.696799] env[62552]: DEBUG nova.compute.manager [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 785.696982] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.698414] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37494239-d763-4717-99eb-d289f955f4ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.708099] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 785.708386] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f1354fe-8a27-4141-a72c-b01f43731252 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.715438] env[62552]: DEBUG oslo_vmware.api [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 785.715438] env[62552]: value = "task-1239310" [ 785.715438] env[62552]: _type = "Task" [ 785.715438] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.726852] env[62552]: DEBUG oslo_vmware.api [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.733969] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.735318] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.271s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.738192] env[62552]: INFO nova.compute.claims [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.753206] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239308, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.082515] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239309, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065126} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.082515] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.082515] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97ced71-985e-4229-bab8-483103cb4209 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.106605] env[62552]: DEBUG nova.compute.manager [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 786.113032] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 00b06df7-e7ae-48fd-8887-65647931d5cf/00b06df7-e7ae-48fd-8887-65647931d5cf.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.115026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.115026] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e37ef1f7-2370-4645-8694-a54dc25147ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.144594] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 786.144594] env[62552]: value = "task-1239311" [ 786.144594] env[62552]: _type = "Task" [ 786.144594] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.153341] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239311, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.229077] env[62552]: DEBUG oslo_vmware.api [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239310, 'name': PowerOffVM_Task, 'duration_secs': 0.199364} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.229708] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.229803] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 786.230782] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3bdec110-ed5f-4d40-a576-357467206d9e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.242875] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquiring lock "48f4f3bc-d957-4ce4-aab6-c32faf0f1113" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.243372] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "48f4f3bc-d957-4ce4-aab6-c32faf0f1113" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.256635] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 786.256920] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 786.257171] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Deleting the datastore file [datastore2] 1472ca01-ba04-4e7d-a9cf-ccaf29c04282 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.257835] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a61dd649-3d3f-4617-8276-89a60f2de231 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.263351] env[62552]: DEBUG oslo_vmware.api [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239308, 'name': PowerOnVM_Task, 'duration_secs': 0.698392} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.265231] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 786.265515] env[62552]: INFO nova.compute.manager [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Took 9.50 seconds to spawn the instance on the hypervisor. [ 786.265746] env[62552]: DEBUG nova.compute.manager [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 786.266189] env[62552]: DEBUG oslo_vmware.api [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for the task: (returnval){ [ 786.266189] env[62552]: value = "task-1239313" [ 786.266189] env[62552]: _type = "Task" [ 786.266189] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.267466] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36681124-9419-4194-99a0-43021ed4d04c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.281296] env[62552]: DEBUG oslo_vmware.api [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239313, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.632647] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.659825] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239311, 'name': ReconfigVM_Task, 'duration_secs': 0.28705} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.660415] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 00b06df7-e7ae-48fd-8887-65647931d5cf/00b06df7-e7ae-48fd-8887-65647931d5cf.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.662299] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-083a31ee-cb12-4688-afe5-df466a990aa0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.669112] env[62552]: INFO nova.scheduler.client.report [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Deleted allocations for instance 1252f45a-d55d-44c0-8629-aba51fb89956 [ 786.677260] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 786.677260] env[62552]: value = "task-1239314" [ 786.677260] env[62552]: _type = "Task" [ 786.677260] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.687563] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239314, 'name': Rename_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.748286] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "48f4f3bc-d957-4ce4-aab6-c32faf0f1113" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.748933] env[62552]: DEBUG nova.compute.manager [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 786.782907] env[62552]: DEBUG oslo_vmware.api [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Task: {'id': task-1239313, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274132} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.786057] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.786167] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 786.786321] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 786.786518] env[62552]: INFO nova.compute.manager [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Took 1.09 seconds to destroy the instance on the hypervisor. [ 786.786845] env[62552]: DEBUG oslo.service.loopingcall [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.794935] env[62552]: DEBUG nova.compute.manager [-] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 786.794935] env[62552]: DEBUG nova.network.neutron [-] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.797288] env[62552]: INFO nova.compute.manager [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Took 31.00 seconds to build instance. [ 786.846325] env[62552]: DEBUG nova.network.neutron [-] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.147879] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9f6717-9b83-4979-a10d-a95e8dc05fad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.156396] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486a44a9-a94c-4718-8020-599f384980dc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.193185] env[62552]: DEBUG oslo_concurrency.lockutils [None req-805f47b4-4ede-49d8-9e7a-724c7b240126 tempest-ServersTestManualDisk-753330648 tempest-ServersTestManualDisk-753330648-project-member] Lock "1252f45a-d55d-44c0-8629-aba51fb89956" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 171.414s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.197389] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4082a5c9-0dbe-4a5f-9984-97d38d74752d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.213106] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239314, 'name': Rename_Task, 'duration_secs': 0.133759} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.213471] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.214767] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86987772-b484-49dc-94fa-3646ce4b38ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.218799] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f93a1a79-fc0c-4158-878e-37110bcab7ee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.232886] env[62552]: DEBUG nova.compute.provider_tree [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.234821] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 787.234821] env[62552]: value = "task-1239315" [ 787.234821] env[62552]: _type = "Task" [ 787.234821] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.248029] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239315, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.254313] env[62552]: DEBUG nova.compute.utils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.256340] env[62552]: DEBUG nova.compute.manager [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 787.256612] env[62552]: DEBUG nova.network.neutron [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 787.299568] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7d0d098d-9ba6-4292-a3ab-62363ea651f2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.898s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.350435] env[62552]: DEBUG nova.network.neutron [-] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.394360] env[62552]: DEBUG nova.policy [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a35514e17a6466aa823c988f97c9aa8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c95e325e75364baf8bad5d28cb9df77d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.705768] env[62552]: DEBUG nova.compute.manager [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 787.737057] env[62552]: DEBUG nova.scheduler.client.report [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 787.754571] env[62552]: DEBUG nova.compute.manager [req-614ca463-d37f-481e-b8e2-09e5ca1be8b9 req-c8937938-02b7-4e89-9c9b-c2a85b8979c9 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Received event network-changed-5cb2de18-f6f3-4351-a156-891bb4a6c542 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 787.754804] env[62552]: DEBUG nova.compute.manager [req-614ca463-d37f-481e-b8e2-09e5ca1be8b9 req-c8937938-02b7-4e89-9c9b-c2a85b8979c9 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Refreshing instance network info cache due to event network-changed-5cb2de18-f6f3-4351-a156-891bb4a6c542. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 787.754985] env[62552]: DEBUG oslo_concurrency.lockutils [req-614ca463-d37f-481e-b8e2-09e5ca1be8b9 req-c8937938-02b7-4e89-9c9b-c2a85b8979c9 service nova] Acquiring lock "refresh_cache-ce5d0165-65f1-4505-9c46-1129c56a8913" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.755144] env[62552]: DEBUG oslo_concurrency.lockutils [req-614ca463-d37f-481e-b8e2-09e5ca1be8b9 req-c8937938-02b7-4e89-9c9b-c2a85b8979c9 service nova] Acquired lock "refresh_cache-ce5d0165-65f1-4505-9c46-1129c56a8913" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.756126] env[62552]: DEBUG nova.network.neutron [req-614ca463-d37f-481e-b8e2-09e5ca1be8b9 req-c8937938-02b7-4e89-9c9b-c2a85b8979c9 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Refreshing network info cache for port 5cb2de18-f6f3-4351-a156-891bb4a6c542 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 787.765347] env[62552]: DEBUG nova.compute.manager [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 787.770035] env[62552]: DEBUG oslo_vmware.api [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239315, 'name': PowerOnVM_Task, 'duration_secs': 0.435986} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.770453] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 787.771805] env[62552]: INFO nova.compute.manager [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Took 8.43 seconds to spawn the instance on the hypervisor. [ 787.772080] env[62552]: DEBUG nova.compute.manager [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 787.773168] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98dcce92-e43b-40c2-a4c7-e01760763935 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.803205] env[62552]: DEBUG nova.compute.manager [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 787.852673] env[62552]: INFO nova.compute.manager [-] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Took 1.06 seconds to deallocate network for instance. [ 788.246150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.258357] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.258908] env[62552]: DEBUG nova.compute.manager [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 788.262852] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.752s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.266712] env[62552]: INFO nova.compute.claims [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.296110] env[62552]: INFO nova.compute.manager [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Took 30.48 seconds to build instance. [ 788.330323] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.357044] env[62552]: DEBUG nova.network.neutron [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Successfully created port: 5ca9cef2-ed2e-436e-a641-e24a499e1714 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.360710] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.384594] env[62552]: DEBUG nova.compute.manager [req-e6ee2fea-34f9-4577-9573-3d152f8423ae req-216e7202-3f63-41bf-9936-34213849a454 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Received event network-changed-abe02bcf-41b7-4884-8ef3-70b5268480f0 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 788.384785] env[62552]: DEBUG nova.compute.manager [req-e6ee2fea-34f9-4577-9573-3d152f8423ae req-216e7202-3f63-41bf-9936-34213849a454 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Refreshing instance network info cache due to event network-changed-abe02bcf-41b7-4884-8ef3-70b5268480f0. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 788.384995] env[62552]: DEBUG oslo_concurrency.lockutils [req-e6ee2fea-34f9-4577-9573-3d152f8423ae req-216e7202-3f63-41bf-9936-34213849a454 service nova] Acquiring lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.385161] env[62552]: DEBUG oslo_concurrency.lockutils [req-e6ee2fea-34f9-4577-9573-3d152f8423ae req-216e7202-3f63-41bf-9936-34213849a454 service nova] Acquired lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.385320] env[62552]: DEBUG nova.network.neutron [req-e6ee2fea-34f9-4577-9573-3d152f8423ae req-216e7202-3f63-41bf-9936-34213849a454 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Refreshing network info cache for port abe02bcf-41b7-4884-8ef3-70b5268480f0 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 788.766088] env[62552]: DEBUG nova.compute.utils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.766088] env[62552]: DEBUG nova.compute.manager [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 788.774018] env[62552]: DEBUG nova.network.neutron [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 788.780918] env[62552]: DEBUG nova.compute.manager [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 788.800304] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0a27538c-c1f0-41b7-afb7-e69ce30fe143 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.467s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.820995] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.821305] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.821533] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.821660] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.821802] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.821943] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.822641] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.822778] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.822995] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.823198] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.823462] env[62552]: DEBUG nova.virt.hardware [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.824627] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f68491a-960e-4f65-a877-d13ef70e95e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.834352] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042d1676-33e6-4e8b-a6a5-efcbcceae7aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.960561] env[62552]: DEBUG nova.policy [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f35eebe4ee3c4c36adf333c8767d0553', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c4aad184cf448d4be4c03e9e6f3e286', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.978283] env[62552]: DEBUG nova.network.neutron [req-614ca463-d37f-481e-b8e2-09e5ca1be8b9 req-c8937938-02b7-4e89-9c9b-c2a85b8979c9 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Updated VIF entry in instance network info cache for port 5cb2de18-f6f3-4351-a156-891bb4a6c542. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 788.978283] env[62552]: DEBUG nova.network.neutron [req-614ca463-d37f-481e-b8e2-09e5ca1be8b9 req-c8937938-02b7-4e89-9c9b-c2a85b8979c9 service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Updating instance_info_cache with network_info: [{"id": "5cb2de18-f6f3-4351-a156-891bb4a6c542", "address": "fa:16:3e:6a:77:08", "network": {"id": "6b678d9d-265f-4541-a994-c70b2ba1774b", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-221581618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2fe0e0c574e4ff2a2ed4acdd4071e42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cb2de18-f6", "ovs_interfaceid": "5cb2de18-f6f3-4351-a156-891bb4a6c542", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.274569] env[62552]: DEBUG nova.compute.manager [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 789.303459] env[62552]: DEBUG nova.compute.manager [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 789.484102] env[62552]: DEBUG oslo_concurrency.lockutils [req-614ca463-d37f-481e-b8e2-09e5ca1be8b9 req-c8937938-02b7-4e89-9c9b-c2a85b8979c9 service nova] Releasing lock "refresh_cache-ce5d0165-65f1-4505-9c46-1129c56a8913" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.492854] env[62552]: DEBUG nova.network.neutron [req-e6ee2fea-34f9-4577-9573-3d152f8423ae req-216e7202-3f63-41bf-9936-34213849a454 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updated VIF entry in instance network info cache for port abe02bcf-41b7-4884-8ef3-70b5268480f0. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 789.493222] env[62552]: DEBUG nova.network.neutron [req-e6ee2fea-34f9-4577-9573-3d152f8423ae req-216e7202-3f63-41bf-9936-34213849a454 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updating instance_info_cache with network_info: [{"id": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "address": "fa:16:3e:82:10:f7", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabe02bcf-41", "ovs_interfaceid": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.673707] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91c6266-76b7-4485-9754-60f79573ec0c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.682968] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c620881-66e5-4589-b333-3f959b51902b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.712273] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eeb2fe3-1d9a-405d-9147-176f002f7257 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.716867] env[62552]: DEBUG nova.network.neutron [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Successfully created port: e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.721965] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd491da-f00e-4328-9910-7e90d79b3ce6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.734993] env[62552]: DEBUG nova.compute.provider_tree [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.822454] env[62552]: DEBUG oslo_concurrency.lockutils [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "00b06df7-e7ae-48fd-8887-65647931d5cf" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.822757] env[62552]: DEBUG oslo_concurrency.lockutils [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.823028] env[62552]: INFO nova.compute.manager [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Rebooting instance [ 789.835472] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.998979] env[62552]: DEBUG oslo_concurrency.lockutils [req-e6ee2fea-34f9-4577-9573-3d152f8423ae req-216e7202-3f63-41bf-9936-34213849a454 service nova] Releasing lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.241498] env[62552]: DEBUG nova.scheduler.client.report [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 790.294613] env[62552]: DEBUG nova.compute.manager [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 790.334061] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.334061] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.334061] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.334337] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.334337] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.334337] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.334337] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.334337] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.335044] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.335387] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.335595] env[62552]: DEBUG nova.virt.hardware [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.339274] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13387d9c-9743-4f2a-a876-8af654ab17d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.348394] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84dacdd-e85e-449c-9383-5ac56fb40d78 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.367828] env[62552]: DEBUG oslo_concurrency.lockutils [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "refresh_cache-00b06df7-e7ae-48fd-8887-65647931d5cf" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.368022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquired lock "refresh_cache-00b06df7-e7ae-48fd-8887-65647931d5cf" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.368258] env[62552]: DEBUG nova.network.neutron [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.470523] env[62552]: DEBUG nova.compute.manager [req-ebb856ef-82ac-4140-8bc5-47d71496bead req-36ad7010-b499-4c76-a776-50e133053995 service nova] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Received event network-vif-plugged-5ca9cef2-ed2e-436e-a641-e24a499e1714 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 790.471726] env[62552]: DEBUG oslo_concurrency.lockutils [req-ebb856ef-82ac-4140-8bc5-47d71496bead req-36ad7010-b499-4c76-a776-50e133053995 service nova] Acquiring lock "4b80bc41-1b63-444e-8039-696e3d78d9a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.471726] env[62552]: DEBUG oslo_concurrency.lockutils [req-ebb856ef-82ac-4140-8bc5-47d71496bead req-36ad7010-b499-4c76-a776-50e133053995 service nova] Lock "4b80bc41-1b63-444e-8039-696e3d78d9a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.471726] env[62552]: DEBUG oslo_concurrency.lockutils [req-ebb856ef-82ac-4140-8bc5-47d71496bead req-36ad7010-b499-4c76-a776-50e133053995 service nova] Lock "4b80bc41-1b63-444e-8039-696e3d78d9a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.471726] env[62552]: DEBUG nova.compute.manager [req-ebb856ef-82ac-4140-8bc5-47d71496bead req-36ad7010-b499-4c76-a776-50e133053995 service nova] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] No waiting events found dispatching network-vif-plugged-5ca9cef2-ed2e-436e-a641-e24a499e1714 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 790.471726] env[62552]: WARNING nova.compute.manager [req-ebb856ef-82ac-4140-8bc5-47d71496bead req-36ad7010-b499-4c76-a776-50e133053995 service nova] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Received unexpected event network-vif-plugged-5ca9cef2-ed2e-436e-a641-e24a499e1714 for instance with vm_state building and task_state spawning. [ 790.748824] env[62552]: DEBUG nova.network.neutron [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Successfully updated port: 5ca9cef2-ed2e-436e-a641-e24a499e1714 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.748824] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.748824] env[62552]: DEBUG nova.compute.manager [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 790.755190] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.783s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.757837] env[62552]: INFO nova.compute.claims [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.258046] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquiring lock "refresh_cache-4b80bc41-1b63-444e-8039-696e3d78d9a1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.258046] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquired lock "refresh_cache-4b80bc41-1b63-444e-8039-696e3d78d9a1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.258046] env[62552]: DEBUG nova.network.neutron [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.270552] env[62552]: DEBUG nova.compute.utils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.277027] env[62552]: DEBUG nova.compute.manager [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 791.277027] env[62552]: DEBUG nova.network.neutron [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 791.327121] env[62552]: DEBUG nova.policy [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d904ba8cf284314baf512c7ea33554b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1716b64900f44a8fb2282801905efd67', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 791.523415] env[62552]: DEBUG nova.network.neutron [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Updating instance_info_cache with network_info: [{"id": "7378d663-07fd-4412-9093-8774fdf7c753", "address": "fa:16:3e:4c:d7:af", "network": {"id": "2bf558f3-1ebd-4738-8e1f-10081487cda2", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1268991274-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ff3ed083bc42bf82311bddcd32d093", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7378d663-07", "ovs_interfaceid": "7378d663-07fd-4412-9093-8774fdf7c753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.621064] env[62552]: DEBUG nova.network.neutron [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Successfully created port: 71e835ff-cc35-4e84-923d-1931fdb1c527 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.782759] env[62552]: DEBUG nova.compute.manager [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 791.992133] env[62552]: DEBUG nova.network.neutron [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.027850] env[62552]: DEBUG oslo_concurrency.lockutils [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Releasing lock "refresh_cache-00b06df7-e7ae-48fd-8887-65647931d5cf" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.137215] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc39debd-11c3-48c2-9056-50356b8dc1da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.147161] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ead47f-bafd-4187-8439-56db45592a60 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.184020] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458f7ad8-8b9b-4f0b-b4b6-ec11eaaafcfb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.191851] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0ee4a5-77e7-4f14-8f3c-8e728d2342e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.205886] env[62552]: DEBUG nova.compute.provider_tree [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.232384] env[62552]: DEBUG nova.network.neutron [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Successfully updated port: e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 792.393388] env[62552]: DEBUG nova.network.neutron [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Updating instance_info_cache with network_info: [{"id": "5ca9cef2-ed2e-436e-a641-e24a499e1714", "address": "fa:16:3e:09:18:c7", "network": {"id": "9863f258-c6da-4ede-bcac-2950f46abcec", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1036448152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95e325e75364baf8bad5d28cb9df77d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ca9cef2-ed", "ovs_interfaceid": "5ca9cef2-ed2e-436e-a641-e24a499e1714", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.531778] env[62552]: DEBUG nova.compute.manager [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 792.532688] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a59c9d1-e96f-4c86-8855-caa9c1d2dea2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.709024] env[62552]: DEBUG nova.scheduler.client.report [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 792.736456] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.736456] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquired lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.736456] env[62552]: DEBUG nova.network.neutron [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.798829] env[62552]: DEBUG nova.compute.manager [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 792.831256] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.831507] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.831661] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.831840] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.831984] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.832175] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.832391] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.832553] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.832717] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.832876] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.833058] env[62552]: DEBUG nova.virt.hardware [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.833925] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ff161a-9fc5-411f-9836-77d36f38017c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.841765] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7b34da-e26b-478b-b2f2-8380decaeb21 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.878728] env[62552]: DEBUG nova.compute.manager [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Received event network-changed-5ca9cef2-ed2e-436e-a641-e24a499e1714 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 792.878920] env[62552]: DEBUG nova.compute.manager [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Refreshing instance network info cache due to event network-changed-5ca9cef2-ed2e-436e-a641-e24a499e1714. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 792.879126] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] Acquiring lock "refresh_cache-4b80bc41-1b63-444e-8039-696e3d78d9a1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.896289] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Releasing lock "refresh_cache-4b80bc41-1b63-444e-8039-696e3d78d9a1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.896601] env[62552]: DEBUG nova.compute.manager [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Instance network_info: |[{"id": "5ca9cef2-ed2e-436e-a641-e24a499e1714", "address": "fa:16:3e:09:18:c7", "network": {"id": "9863f258-c6da-4ede-bcac-2950f46abcec", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1036448152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95e325e75364baf8bad5d28cb9df77d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ca9cef2-ed", "ovs_interfaceid": "5ca9cef2-ed2e-436e-a641-e24a499e1714", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 792.897023] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] Acquired lock "refresh_cache-4b80bc41-1b63-444e-8039-696e3d78d9a1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.897260] env[62552]: DEBUG nova.network.neutron [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Refreshing network info cache for port 5ca9cef2-ed2e-436e-a641-e24a499e1714 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 792.898450] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:18:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ca9cef2-ed2e-436e-a641-e24a499e1714', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.906097] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Creating folder: Project (c95e325e75364baf8bad5d28cb9df77d). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.908024] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1512dd9-61ce-4f93-90c2-801354b9adf1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.916968] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Created folder: Project (c95e325e75364baf8bad5d28cb9df77d) in parent group-v267339. [ 792.917144] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Creating folder: Instances. Parent ref: group-v267377. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.917576] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e37ec906-ad3f-4b5a-9670-187fb0705c63 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.926903] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Created folder: Instances in parent group-v267377. [ 792.929792] env[62552]: DEBUG oslo.service.loopingcall [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.929792] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.929792] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b7e25b0-d2ae-4284-ba97-7cf823914cb9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.948895] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.948895] env[62552]: value = "task-1239318" [ 792.948895] env[62552]: _type = "Task" [ 792.948895] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.957290] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239318, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.058638] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "fb881c34-dccc-4703-af8e-c75caafd9b08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.058882] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fb881c34-dccc-4703-af8e-c75caafd9b08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.146286] env[62552]: DEBUG nova.compute.manager [req-881db2d9-8f43-426b-be24-cddb0ea98241 req-5fcf4117-d9f0-4321-99e5-387c22421f86 service nova] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Received event network-vif-plugged-71e835ff-cc35-4e84-923d-1931fdb1c527 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 793.146784] env[62552]: DEBUG oslo_concurrency.lockutils [req-881db2d9-8f43-426b-be24-cddb0ea98241 req-5fcf4117-d9f0-4321-99e5-387c22421f86 service nova] Acquiring lock "714fb65d-9f80-4a81-a637-3e4398405d9b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.146784] env[62552]: DEBUG oslo_concurrency.lockutils [req-881db2d9-8f43-426b-be24-cddb0ea98241 req-5fcf4117-d9f0-4321-99e5-387c22421f86 service nova] Lock "714fb65d-9f80-4a81-a637-3e4398405d9b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.147138] env[62552]: DEBUG oslo_concurrency.lockutils [req-881db2d9-8f43-426b-be24-cddb0ea98241 req-5fcf4117-d9f0-4321-99e5-387c22421f86 service nova] Lock "714fb65d-9f80-4a81-a637-3e4398405d9b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.147138] env[62552]: DEBUG nova.compute.manager [req-881db2d9-8f43-426b-be24-cddb0ea98241 req-5fcf4117-d9f0-4321-99e5-387c22421f86 service nova] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] No waiting events found dispatching network-vif-plugged-71e835ff-cc35-4e84-923d-1931fdb1c527 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 793.147234] env[62552]: WARNING nova.compute.manager [req-881db2d9-8f43-426b-be24-cddb0ea98241 req-5fcf4117-d9f0-4321-99e5-387c22421f86 service nova] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Received unexpected event network-vif-plugged-71e835ff-cc35-4e84-923d-1931fdb1c527 for instance with vm_state building and task_state spawning. [ 793.223362] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.223362] env[62552]: DEBUG nova.compute.manager [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 793.231029] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.930s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.231313] env[62552]: DEBUG nova.objects.instance [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62552) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 793.281246] env[62552]: DEBUG nova.network.neutron [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Successfully updated port: 71e835ff-cc35-4e84-923d-1931fdb1c527 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.336305] env[62552]: DEBUG nova.network.neutron [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.461965] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239318, 'name': CreateVM_Task, 'duration_secs': 0.343727} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.463109] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 793.463109] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.463283] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.463517] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.463784] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec5c906a-f933-4a1c-9d68-d0993b62e8ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.469046] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 793.469046] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a35e51-cff4-d493-8bc0-167e9bba9013" [ 793.469046] env[62552]: _type = "Task" [ 793.469046] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.485973] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a35e51-cff4-d493-8bc0-167e9bba9013, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.551302] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35cebc9c-077c-415a-8167-9e5ec2a14775 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.563836] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Doing hard reboot of VM {{(pid=62552) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 793.563836] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-beec2de4-ba4d-4b84-b259-29bf68365dea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.570756] env[62552]: DEBUG oslo_vmware.api [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 793.570756] env[62552]: value = "task-1239319" [ 793.570756] env[62552]: _type = "Task" [ 793.570756] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.578820] env[62552]: DEBUG oslo_vmware.api [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239319, 'name': ResetVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.644448] env[62552]: DEBUG nova.network.neutron [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Updated VIF entry in instance network info cache for port 5ca9cef2-ed2e-436e-a641-e24a499e1714. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 793.645735] env[62552]: DEBUG nova.network.neutron [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Updating instance_info_cache with network_info: [{"id": "5ca9cef2-ed2e-436e-a641-e24a499e1714", "address": "fa:16:3e:09:18:c7", "network": {"id": "9863f258-c6da-4ede-bcac-2950f46abcec", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1036448152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95e325e75364baf8bad5d28cb9df77d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ca9cef2-ed", "ovs_interfaceid": "5ca9cef2-ed2e-436e-a641-e24a499e1714", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.691418] env[62552]: DEBUG nova.network.neutron [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updating instance_info_cache with network_info: [{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.699286] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "fea83dde-3181-49dd-a000-5e5cffc8de95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.699286] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fea83dde-3181-49dd-a000-5e5cffc8de95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.737514] env[62552]: DEBUG nova.compute.utils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.747018] env[62552]: DEBUG nova.compute.manager [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 793.747018] env[62552]: DEBUG nova.network.neutron [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 793.788061] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "refresh_cache-714fb65d-9f80-4a81-a637-3e4398405d9b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.788254] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "refresh_cache-714fb65d-9f80-4a81-a637-3e4398405d9b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.788870] env[62552]: DEBUG nova.network.neutron [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.841273] env[62552]: DEBUG nova.policy [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '384d6877b7384d21b12a155d13056a7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f13a11803d1b44178f053cc0a4affedd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 793.979968] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a35e51-cff4-d493-8bc0-167e9bba9013, 'name': SearchDatastore_Task, 'duration_secs': 0.011019} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.980310] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.980544] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.980772] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.980913] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.981099] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.981375] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da6a64bd-73d0-429d-bd0d-f91f34690913 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.988910] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.989097] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 793.990049] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdf00251-e738-45d9-827d-e8dc61d6b218 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.995076] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 793.995076] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5281f321-8300-88dc-c0f1-93bc7e5abe60" [ 793.995076] env[62552]: _type = "Task" [ 793.995076] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.002858] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5281f321-8300-88dc-c0f1-93bc7e5abe60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.081018] env[62552]: DEBUG oslo_vmware.api [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239319, 'name': ResetVM_Task, 'duration_secs': 0.123435} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.085042] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Did hard reboot of VM {{(pid=62552) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 794.085042] env[62552]: DEBUG nova.compute.manager [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 794.085042] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a517a7-1d30-485c-af6a-924f94c80e11 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.151021] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] Releasing lock "refresh_cache-4b80bc41-1b63-444e-8039-696e3d78d9a1" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.151021] env[62552]: DEBUG nova.compute.manager [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Received event network-vif-plugged-e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 794.151021] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] Acquiring lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.151021] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] Lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.151021] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] Lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.151275] env[62552]: DEBUG nova.compute.manager [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] No waiting events found dispatching network-vif-plugged-e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 794.151275] env[62552]: WARNING nova.compute.manager [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Received unexpected event network-vif-plugged-e3439c52-55fc-4ea8-ae30-607c154031cb for instance with vm_state building and task_state spawning. [ 794.151275] env[62552]: DEBUG nova.compute.manager [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Received event network-changed-e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 794.151275] env[62552]: DEBUG nova.compute.manager [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Refreshing instance network info cache due to event network-changed-e3439c52-55fc-4ea8-ae30-607c154031cb. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 794.151275] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] Acquiring lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.194655] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Releasing lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.194978] env[62552]: DEBUG nova.compute.manager [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Instance network_info: |[{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 794.195270] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] Acquired lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.195845] env[62552]: DEBUG nova.network.neutron [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Refreshing network info cache for port e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.196783] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:99:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff1f3320-df8e-49df-a412-9797a23bd173', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e3439c52-55fc-4ea8-ae30-607c154031cb', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.207093] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Creating folder: Project (3c4aad184cf448d4be4c03e9e6f3e286). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.211120] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d92dfe0c-ae41-41bd-b86a-f6377de2a6cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.224137] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Created folder: Project (3c4aad184cf448d4be4c03e9e6f3e286) in parent group-v267339. [ 794.224137] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Creating folder: Instances. Parent ref: group-v267380. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.224137] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd66ca18-c29a-41ef-981f-450ad64275d4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.234449] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Created folder: Instances in parent group-v267380. [ 794.234449] env[62552]: DEBUG oslo.service.loopingcall [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.234449] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 794.234449] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-788a3a01-2119-4908-96bc-772903b968ca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.254056] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c93d8112-61af-4fa0-8a49-75d653ab19c1 tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.030s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.255855] env[62552]: DEBUG nova.compute.manager [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 794.260091] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.939s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.264499] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.264499] env[62552]: value = "task-1239322" [ 794.264499] env[62552]: _type = "Task" [ 794.264499] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.272166] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239322, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.374826] env[62552]: DEBUG nova.network.neutron [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.442824] env[62552]: DEBUG nova.network.neutron [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updated VIF entry in instance network info cache for port e3439c52-55fc-4ea8-ae30-607c154031cb. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 794.443752] env[62552]: DEBUG nova.network.neutron [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updating instance_info_cache with network_info: [{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.513382] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5281f321-8300-88dc-c0f1-93bc7e5abe60, 'name': SearchDatastore_Task, 'duration_secs': 0.010344} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.515472] env[62552]: DEBUG nova.network.neutron [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Successfully created port: 1e51c71e-8255-4974-a13e-53631ef167af {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.518463] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b612a9d-1a7c-414b-a30a-a2c1f9968789 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.526581] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 794.526581] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5290a1ff-593f-dcad-227d-3866091bae1c" [ 794.526581] env[62552]: _type = "Task" [ 794.526581] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.541795] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5290a1ff-593f-dcad-227d-3866091bae1c, 'name': SearchDatastore_Task, 'duration_secs': 0.009091} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.542162] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.542499] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 4b80bc41-1b63-444e-8039-696e3d78d9a1/4b80bc41-1b63-444e-8039-696e3d78d9a1.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 794.542888] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-747afcfc-4307-43e9-b580-d6e4de7fc2a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.551452] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 794.551452] env[62552]: value = "task-1239323" [ 794.551452] env[62552]: _type = "Task" [ 794.551452] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.561126] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.595577] env[62552]: DEBUG oslo_concurrency.lockutils [None req-165aafe3-2df1-4e66-8c2b-889f09616ee0 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.773s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.584756] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9f398ff-d604-4118-ab95-674dd2935314 req-091f217e-55f2-4444-8479-a6d855b3f38d service nova] Releasing lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.586071] env[62552]: DEBUG nova.compute.manager [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 795.588495] env[62552]: DEBUG nova.compute.manager [req-0f731661-88bb-4499-9bde-ef6129e74a16 req-63f591ed-8725-4f53-8ef6-298c0ae6d9d9 service nova] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Received event network-changed-71e835ff-cc35-4e84-923d-1931fdb1c527 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 795.588675] env[62552]: DEBUG nova.compute.manager [req-0f731661-88bb-4499-9bde-ef6129e74a16 req-63f591ed-8725-4f53-8ef6-298c0ae6d9d9 service nova] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Refreshing instance network info cache due to event network-changed-71e835ff-cc35-4e84-923d-1931fdb1c527. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 795.588856] env[62552]: DEBUG oslo_concurrency.lockutils [req-0f731661-88bb-4499-9bde-ef6129e74a16 req-63f591ed-8725-4f53-8ef6-298c0ae6d9d9 service nova] Acquiring lock "refresh_cache-714fb65d-9f80-4a81-a637-3e4398405d9b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.600199] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239322, 'name': CreateVM_Task, 'duration_secs': 0.406348} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.603638] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 795.604015] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239323, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514491} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.604667] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.604855] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.605124] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.605452] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 4b80bc41-1b63-444e-8039-696e3d78d9a1/4b80bc41-1b63-444e-8039-696e3d78d9a1.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 795.605568] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.605867] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3924b67-e1da-44c6-9a5d-a94fb2b4b898 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.607667] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dee6d549-d644-41a3-866a-b734d449daee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.612788] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 795.612788] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521602bd-ea3a-db93-63e1-3d1e311e7b0e" [ 795.612788] env[62552]: _type = "Task" [ 795.612788] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.619055] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 795.619055] env[62552]: value = "task-1239324" [ 795.619055] env[62552]: _type = "Task" [ 795.619055] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.620377] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance cdee22a9-4327-47af-ab14-8403a06b2802 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.620538] env[62552]: WARNING nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 1472ca01-ba04-4e7d-a9cf-ccaf29c04282 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 795.620580] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 8707ce4b-677e-4f13-86f8-3e327d19380b actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.620702] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance ce5d0165-65f1-4505-9c46-1129c56a8913 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.620825] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 0418260a-aa27-4955-ab15-b180ec04f0b3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.620941] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 00b06df7-e7ae-48fd-8887-65647931d5cf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.621072] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 4b80bc41-1b63-444e-8039-696e3d78d9a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.621174] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 51bbcc4e-8251-4b38-9d36-8aea8fc7705d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.621285] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 714fb65d-9f80-4a81-a637-3e4398405d9b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.621395] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 34736dd0-e617-475e-baa2-cb372db1afb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.629373] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.629373] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.629373] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.629373] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.629576] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.629576] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.629971] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.629971] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.629971] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.630174] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.630528] env[62552]: DEBUG nova.virt.hardware [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.634430] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb30cde-b660-4fb7-96fd-d5aa06f39d5a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.636909] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521602bd-ea3a-db93-63e1-3d1e311e7b0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.642019] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239324, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.645861] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ae7512-02cc-4f5d-a7c5-960a48b58c43 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.662148] env[62552]: DEBUG nova.network.neutron [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Updating instance_info_cache with network_info: [{"id": "71e835ff-cc35-4e84-923d-1931fdb1c527", "address": "fa:16:3e:61:46:82", "network": {"id": "925fdab2-f263-46f3-ab17-2f51407bae3e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1178354148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1716b64900f44a8fb2282801905efd67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e835ff-cc", "ovs_interfaceid": "71e835ff-cc35-4e84-923d-1931fdb1c527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.126318] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521602bd-ea3a-db93-63e1-3d1e311e7b0e, 'name': SearchDatastore_Task, 'duration_secs': 0.0139} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.129506] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.130020] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.130020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.130297] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.130436] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.131157] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 36233c83-1170-4a7a-be0b-6a0e8b139de7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.132609] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63630246-965e-4bb2-9a88-f50abac9f857 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.149120] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239324, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118983} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.152710] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.152710] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.152710] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 796.152710] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e043a6-5b11-46a5-9ea2-1586b58cb621 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.156486] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1699ea7-1a30-4df1-8fa5-0d96f81316ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.171693] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "refresh_cache-714fb65d-9f80-4a81-a637-3e4398405d9b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.172024] env[62552]: DEBUG nova.compute.manager [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Instance network_info: |[{"id": "71e835ff-cc35-4e84-923d-1931fdb1c527", "address": "fa:16:3e:61:46:82", "network": {"id": "925fdab2-f263-46f3-ab17-2f51407bae3e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1178354148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1716b64900f44a8fb2282801905efd67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e835ff-cc", "ovs_interfaceid": "71e835ff-cc35-4e84-923d-1931fdb1c527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 796.183349] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 4b80bc41-1b63-444e-8039-696e3d78d9a1/4b80bc41-1b63-444e-8039-696e3d78d9a1.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.185496] env[62552]: DEBUG oslo_concurrency.lockutils [req-0f731661-88bb-4499-9bde-ef6129e74a16 req-63f591ed-8725-4f53-8ef6-298c0ae6d9d9 service nova] Acquired lock "refresh_cache-714fb65d-9f80-4a81-a637-3e4398405d9b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.185496] env[62552]: DEBUG nova.network.neutron [req-0f731661-88bb-4499-9bde-ef6129e74a16 req-63f591ed-8725-4f53-8ef6-298c0ae6d9d9 service nova] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Refreshing network info cache for port 71e835ff-cc35-4e84-923d-1931fdb1c527 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.186716] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:46:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '71e835ff-cc35-4e84-923d-1931fdb1c527', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.195066] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Creating folder: Project (1716b64900f44a8fb2282801905efd67). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.197591] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9466488a-f638-4f10-b4d9-b1ebbfa63949 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.214265] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 796.214265] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52086cfc-7774-4d69-462d-0bdd29fa6e45" [ 796.214265] env[62552]: _type = "Task" [ 796.214265] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.216763] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6724989f-7eec-4085-970c-4889a5fe286b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.225895] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 796.225895] env[62552]: value = "task-1239325" [ 796.225895] env[62552]: _type = "Task" [ 796.225895] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.236789] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52086cfc-7774-4d69-462d-0bdd29fa6e45, 'name': SearchDatastore_Task, 'duration_secs': 0.031517} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.239354] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Created folder: Project (1716b64900f44a8fb2282801905efd67) in parent group-v267339. [ 796.239590] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Creating folder: Instances. Parent ref: group-v267383. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.239872] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f40a6bb5-d13a-4006-a403-d99afdff3c1e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.243409] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ebce67c3-734c-4ba1-a60b-5bb293d70cf4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.248258] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.253474] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 796.253474] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d1fbd1-c492-b288-11e0-9cc75a8ae7d6" [ 796.253474] env[62552]: _type = "Task" [ 796.253474] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.256957] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Created folder: Instances in parent group-v267383. [ 796.257438] env[62552]: DEBUG oslo.service.loopingcall [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.258029] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 796.258325] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cfee6ee8-7947-4783-8513-9d209e797e1c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.279066] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "00b06df7-e7ae-48fd-8887-65647931d5cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.279349] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.279569] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "00b06df7-e7ae-48fd-8887-65647931d5cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.279753] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.279927] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.281891] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d1fbd1-c492-b288-11e0-9cc75a8ae7d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.282412] env[62552]: INFO nova.compute.manager [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Terminating instance [ 796.295681] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.295681] env[62552]: value = "task-1239328" [ 796.295681] env[62552]: _type = "Task" [ 796.295681] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.301914] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239328, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.638825] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 3b333977-0b4b-4ccf-bed5-231215abae3c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.661187] env[62552]: DEBUG nova.network.neutron [req-0f731661-88bb-4499-9bde-ef6129e74a16 req-63f591ed-8725-4f53-8ef6-298c0ae6d9d9 service nova] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Updated VIF entry in instance network info cache for port 71e835ff-cc35-4e84-923d-1931fdb1c527. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 796.661572] env[62552]: DEBUG nova.network.neutron [req-0f731661-88bb-4499-9bde-ef6129e74a16 req-63f591ed-8725-4f53-8ef6-298c0ae6d9d9 service nova] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Updating instance_info_cache with network_info: [{"id": "71e835ff-cc35-4e84-923d-1931fdb1c527", "address": "fa:16:3e:61:46:82", "network": {"id": "925fdab2-f263-46f3-ab17-2f51407bae3e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1178354148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1716b64900f44a8fb2282801905efd67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e835ff-cc", "ovs_interfaceid": "71e835ff-cc35-4e84-923d-1931fdb1c527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.738563] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239325, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.765117] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d1fbd1-c492-b288-11e0-9cc75a8ae7d6, 'name': SearchDatastore_Task, 'duration_secs': 0.027507} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.765117] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.765117] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 51bbcc4e-8251-4b38-9d36-8aea8fc7705d/51bbcc4e-8251-4b38-9d36-8aea8fc7705d.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 796.765117] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32031b42-2b93-44d6-8001-d256cd167f7d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.770267] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 796.770267] env[62552]: value = "task-1239329" [ 796.770267] env[62552]: _type = "Task" [ 796.770267] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.780905] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239329, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.789193] env[62552]: DEBUG nova.compute.manager [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 796.789193] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.789193] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4446243e-f5f2-4f02-8be2-0a05c46d2686 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.798379] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.799154] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ce272f8-bcad-491f-a6f9-2b84fdf44ca5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.804476] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239328, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.810020] env[62552]: DEBUG oslo_vmware.api [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 796.810020] env[62552]: value = "task-1239330" [ 796.810020] env[62552]: _type = "Task" [ 796.810020] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.821568] env[62552]: DEBUG oslo_vmware.api [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.861965] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquiring lock "fb47f136-7859-4756-bc92-fbcf430b4c8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.862483] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "fb47f136-7859-4756-bc92-fbcf430b4c8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.993725] env[62552]: DEBUG nova.compute.manager [req-92104dba-3b22-4301-b1cb-3c4cf2735f26 req-55379f66-1d0d-442a-994d-736dfc95cda3 service nova] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Received event network-vif-plugged-1e51c71e-8255-4974-a13e-53631ef167af {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 796.994131] env[62552]: DEBUG oslo_concurrency.lockutils [req-92104dba-3b22-4301-b1cb-3c4cf2735f26 req-55379f66-1d0d-442a-994d-736dfc95cda3 service nova] Acquiring lock "34736dd0-e617-475e-baa2-cb372db1afb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.994481] env[62552]: DEBUG oslo_concurrency.lockutils [req-92104dba-3b22-4301-b1cb-3c4cf2735f26 req-55379f66-1d0d-442a-994d-736dfc95cda3 service nova] Lock "34736dd0-e617-475e-baa2-cb372db1afb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.994901] env[62552]: DEBUG oslo_concurrency.lockutils [req-92104dba-3b22-4301-b1cb-3c4cf2735f26 req-55379f66-1d0d-442a-994d-736dfc95cda3 service nova] Lock "34736dd0-e617-475e-baa2-cb372db1afb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.995248] env[62552]: DEBUG nova.compute.manager [req-92104dba-3b22-4301-b1cb-3c4cf2735f26 req-55379f66-1d0d-442a-994d-736dfc95cda3 service nova] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] No waiting events found dispatching network-vif-plugged-1e51c71e-8255-4974-a13e-53631ef167af {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 796.995517] env[62552]: WARNING nova.compute.manager [req-92104dba-3b22-4301-b1cb-3c4cf2735f26 req-55379f66-1d0d-442a-994d-736dfc95cda3 service nova] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Received unexpected event network-vif-plugged-1e51c71e-8255-4974-a13e-53631ef167af for instance with vm_state building and task_state spawning. [ 797.076140] env[62552]: DEBUG nova.network.neutron [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Successfully updated port: 1e51c71e-8255-4974-a13e-53631ef167af {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 797.146255] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance fef94e30-f946-4d7f-a108-f32cad505bc4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.164400] env[62552]: DEBUG oslo_concurrency.lockutils [req-0f731661-88bb-4499-9bde-ef6129e74a16 req-63f591ed-8725-4f53-8ef6-298c0ae6d9d9 service nova] Releasing lock "refresh_cache-714fb65d-9f80-4a81-a637-3e4398405d9b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.238542] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239325, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.280873] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239329, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464285} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.281146] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 51bbcc4e-8251-4b38-9d36-8aea8fc7705d/51bbcc4e-8251-4b38-9d36-8aea8fc7705d.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 797.281390] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 797.281644] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-510bfbdc-312a-4baf-9454-bc3b167615b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.288296] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 797.288296] env[62552]: value = "task-1239331" [ 797.288296] env[62552]: _type = "Task" [ 797.288296] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.295436] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239331, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.303023] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239328, 'name': CreateVM_Task, 'duration_secs': 0.524374} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.303023] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 797.303322] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.303474] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.303800] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.303988] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6bce6cf-8b5d-495a-bee5-60619cfe7733 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.308970] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 797.308970] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fed134-4cbf-fa22-46f0-a40d2c8eb828" [ 797.308970] env[62552]: _type = "Task" [ 797.308970] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.319060] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fed134-4cbf-fa22-46f0-a40d2c8eb828, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.321389] env[62552]: DEBUG oslo_vmware.api [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239330, 'name': PowerOffVM_Task, 'duration_secs': 0.214869} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.321506] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.321666] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 797.321890] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a42c153a-6a97-4667-8bd3-fbc3c089634d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.398296] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.398296] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.398296] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Deleting the datastore file [datastore2] 00b06df7-e7ae-48fd-8887-65647931d5cf {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.398296] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-903d4997-3f56-460c-9064-6dfedfd322e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.404948] env[62552]: DEBUG oslo_vmware.api [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for the task: (returnval){ [ 797.404948] env[62552]: value = "task-1239333" [ 797.404948] env[62552]: _type = "Task" [ 797.404948] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.413933] env[62552]: DEBUG oslo_vmware.api [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239333, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.580592] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "refresh_cache-34736dd0-e617-475e-baa2-cb372db1afb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.580592] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "refresh_cache-34736dd0-e617-475e-baa2-cb372db1afb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.580701] env[62552]: DEBUG nova.network.neutron [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 797.649918] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 80478878-ff82-4ed6-a851-8eb2bec01e22 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.738924] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239325, 'name': ReconfigVM_Task, 'duration_secs': 1.455289} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.739219] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 4b80bc41-1b63-444e-8039-696e3d78d9a1/4b80bc41-1b63-444e-8039-696e3d78d9a1.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.739833] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc69b6b2-9d80-4b0e-b563-80b93b00460c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.746479] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 797.746479] env[62552]: value = "task-1239334" [ 797.746479] env[62552]: _type = "Task" [ 797.746479] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.754911] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239334, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.798221] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239331, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.4131} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.798499] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.799253] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60cfba42-8b60-4e1e-b382-c47804f7cb4f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.820419] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 51bbcc4e-8251-4b38-9d36-8aea8fc7705d/51bbcc4e-8251-4b38-9d36-8aea8fc7705d.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.823279] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b34fbeb3-4b06-4d12-9660-231b2e4b154d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.843560] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fed134-4cbf-fa22-46f0-a40d2c8eb828, 'name': SearchDatastore_Task, 'duration_secs': 0.00889} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.844843] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.845094] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.845324] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.845471] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.845642] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.845944] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 797.845944] env[62552]: value = "task-1239335" [ 797.845944] env[62552]: _type = "Task" [ 797.845944] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.846142] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45f96daa-3fd7-414f-9cce-d99a75a82331 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.856476] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239335, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.863855] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.864047] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 797.864732] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa0671de-d0cb-44f2-9c27-77cd0eceb0c7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.869894] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 797.869894] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520137e3-c331-0d93-2fa0-acd3ed255d2f" [ 797.869894] env[62552]: _type = "Task" [ 797.869894] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.877349] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520137e3-c331-0d93-2fa0-acd3ed255d2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.913850] env[62552]: DEBUG oslo_vmware.api [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Task: {'id': task-1239333, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138283} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.914127] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.914324] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 797.914498] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.914668] env[62552]: INFO nova.compute.manager [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Took 1.13 seconds to destroy the instance on the hypervisor. [ 797.914904] env[62552]: DEBUG oslo.service.loopingcall [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.915097] env[62552]: DEBUG nova.compute.manager [-] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 797.915193] env[62552]: DEBUG nova.network.neutron [-] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 798.152240] env[62552]: DEBUG nova.network.neutron [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.154801] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.257145] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239334, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.333581] env[62552]: DEBUG nova.network.neutron [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Updating instance_info_cache with network_info: [{"id": "1e51c71e-8255-4974-a13e-53631ef167af", "address": "fa:16:3e:b9:0e:24", "network": {"id": "6ad33460-d660-442d-a742-61dcc66532b5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1461793279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f13a11803d1b44178f053cc0a4affedd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e51c71e-82", "ovs_interfaceid": "1e51c71e-8255-4974-a13e-53631ef167af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.359017] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239335, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.382659] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520137e3-c331-0d93-2fa0-acd3ed255d2f, 'name': SearchDatastore_Task, 'duration_secs': 0.021352} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.383869] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ac395f1-8f45-49cf-a6a3-2a31068cd238 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.390619] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 798.390619] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525191e7-eb92-57a6-78d8-059121a11020" [ 798.390619] env[62552]: _type = "Task" [ 798.390619] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.400637] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525191e7-eb92-57a6-78d8-059121a11020, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.660159] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 88d0f502-fc3f-429e-bdf6-a1ebacec2117 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.742045] env[62552]: DEBUG nova.network.neutron [-] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.757743] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239334, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.836131] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "refresh_cache-34736dd0-e617-475e-baa2-cb372db1afb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.836760] env[62552]: DEBUG nova.compute.manager [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Instance network_info: |[{"id": "1e51c71e-8255-4974-a13e-53631ef167af", "address": "fa:16:3e:b9:0e:24", "network": {"id": "6ad33460-d660-442d-a742-61dcc66532b5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1461793279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f13a11803d1b44178f053cc0a4affedd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e51c71e-82", "ovs_interfaceid": "1e51c71e-8255-4974-a13e-53631ef167af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 798.836975] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:0e:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e51c71e-8255-4974-a13e-53631ef167af', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.844250] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Creating folder: Project (f13a11803d1b44178f053cc0a4affedd). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.844527] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1c9adbc-68ce-4550-bb65-4bf791bef2f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.854917] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Created folder: Project (f13a11803d1b44178f053cc0a4affedd) in parent group-v267339. [ 798.855046] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Creating folder: Instances. Parent ref: group-v267386. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.858298] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e478562-1d10-4b92-a345-82f0a580fd4d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.859675] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239335, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.866996] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Created folder: Instances in parent group-v267386. [ 798.867237] env[62552]: DEBUG oslo.service.loopingcall [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.867428] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 798.867658] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6be90843-361a-4266-affb-55cf9383b201 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.886873] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.886873] env[62552]: value = "task-1239338" [ 798.886873] env[62552]: _type = "Task" [ 798.886873] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.897071] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239338, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.902091] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525191e7-eb92-57a6-78d8-059121a11020, 'name': SearchDatastore_Task, 'duration_secs': 0.008948} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.902329] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.902603] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 714fb65d-9f80-4a81-a637-3e4398405d9b/714fb65d-9f80-4a81-a637-3e4398405d9b.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 798.902831] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-504e31f8-c731-43ba-8250-d17af2ca37f8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.910202] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 798.910202] env[62552]: value = "task-1239339" [ 798.910202] env[62552]: _type = "Task" [ 798.910202] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.917343] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239339, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.041334] env[62552]: DEBUG nova.compute.manager [req-fea1e64a-14fe-4aa8-be41-0f2fbfe890ed req-f293fddc-2040-44a1-9515-ba290d52013a service nova] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Received event network-changed-1e51c71e-8255-4974-a13e-53631ef167af {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 799.041560] env[62552]: DEBUG nova.compute.manager [req-fea1e64a-14fe-4aa8-be41-0f2fbfe890ed req-f293fddc-2040-44a1-9515-ba290d52013a service nova] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Refreshing instance network info cache due to event network-changed-1e51c71e-8255-4974-a13e-53631ef167af. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 799.041813] env[62552]: DEBUG oslo_concurrency.lockutils [req-fea1e64a-14fe-4aa8-be41-0f2fbfe890ed req-f293fddc-2040-44a1-9515-ba290d52013a service nova] Acquiring lock "refresh_cache-34736dd0-e617-475e-baa2-cb372db1afb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.041986] env[62552]: DEBUG oslo_concurrency.lockutils [req-fea1e64a-14fe-4aa8-be41-0f2fbfe890ed req-f293fddc-2040-44a1-9515-ba290d52013a service nova] Acquired lock "refresh_cache-34736dd0-e617-475e-baa2-cb372db1afb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.042186] env[62552]: DEBUG nova.network.neutron [req-fea1e64a-14fe-4aa8-be41-0f2fbfe890ed req-f293fddc-2040-44a1-9515-ba290d52013a service nova] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Refreshing network info cache for port 1e51c71e-8255-4974-a13e-53631ef167af {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 799.163699] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 474dfa9e-d003-478f-b48f-09e0c0452ffd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.246031] env[62552]: INFO nova.compute.manager [-] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Took 1.33 seconds to deallocate network for instance. [ 799.264642] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239334, 'name': Rename_Task, 'duration_secs': 1.071551} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.264971] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 799.265243] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51cfa022-aa22-4ed7-bf35-fc2b8808fe78 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.273215] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 799.273215] env[62552]: value = "task-1239340" [ 799.273215] env[62552]: _type = "Task" [ 799.273215] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.283534] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239340, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.360335] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239335, 'name': ReconfigVM_Task, 'duration_secs': 1.115584} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.360641] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 51bbcc4e-8251-4b38-9d36-8aea8fc7705d/51bbcc4e-8251-4b38-9d36-8aea8fc7705d.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.361340] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-794e8530-9c1b-4f30-9392-944237afafb9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.367351] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 799.367351] env[62552]: value = "task-1239341" [ 799.367351] env[62552]: _type = "Task" [ 799.367351] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.375030] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239341, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.394926] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239338, 'name': CreateVM_Task, 'duration_secs': 0.288343} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.395096] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 799.395782] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.395938] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.396263] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.396509] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6adb2110-4c48-405f-94fd-56bf98edd1ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.400949] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 799.400949] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5224db2a-afd8-6a89-d1bb-f4c9f9ef7be1" [ 799.400949] env[62552]: _type = "Task" [ 799.400949] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.408390] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5224db2a-afd8-6a89-d1bb-f4c9f9ef7be1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.417537] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239339, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465282} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.417773] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 714fb65d-9f80-4a81-a637-3e4398405d9b/714fb65d-9f80-4a81-a637-3e4398405d9b.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.417975] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.418239] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10d83515-bcae-4907-b5d3-b45233627ce9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.426089] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 799.426089] env[62552]: value = "task-1239342" [ 799.426089] env[62552]: _type = "Task" [ 799.426089] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.433483] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.666800] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance c52efdeb-9efc-4ba1-b9af-4dd0a925517f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.757014] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.784360] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239340, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.817475] env[62552]: DEBUG nova.network.neutron [req-fea1e64a-14fe-4aa8-be41-0f2fbfe890ed req-f293fddc-2040-44a1-9515-ba290d52013a service nova] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Updated VIF entry in instance network info cache for port 1e51c71e-8255-4974-a13e-53631ef167af. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 799.818499] env[62552]: DEBUG nova.network.neutron [req-fea1e64a-14fe-4aa8-be41-0f2fbfe890ed req-f293fddc-2040-44a1-9515-ba290d52013a service nova] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Updating instance_info_cache with network_info: [{"id": "1e51c71e-8255-4974-a13e-53631ef167af", "address": "fa:16:3e:b9:0e:24", "network": {"id": "6ad33460-d660-442d-a742-61dcc66532b5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1461793279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f13a11803d1b44178f053cc0a4affedd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e51c71e-82", "ovs_interfaceid": "1e51c71e-8255-4974-a13e-53631ef167af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.876926] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239341, 'name': Rename_Task, 'duration_secs': 0.163097} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.877285] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 799.877564] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f3560a2-666c-4303-aa5f-3e67d7858bd7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.884395] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 799.884395] env[62552]: value = "task-1239343" [ 799.884395] env[62552]: _type = "Task" [ 799.884395] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.892167] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239343, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.910148] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5224db2a-afd8-6a89-d1bb-f4c9f9ef7be1, 'name': SearchDatastore_Task, 'duration_secs': 0.008133} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.910491] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.910770] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.911057] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.911274] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.911520] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.911806] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f410baa-265a-4753-aa84-50614f717167 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.920008] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.920203] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.920900] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e4344ba-75ec-4b7d-94ef-aba7985a1119 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.925850] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 799.925850] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bc5002-0182-c070-5607-22215dcf47ab" [ 799.925850] env[62552]: _type = "Task" [ 799.925850] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.936980] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bc5002-0182-c070-5607-22215dcf47ab, 'name': SearchDatastore_Task, 'duration_secs': 0.007614} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.940121] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062654} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.940394] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9cbf7ae2-ad68-4f02-ac76-4402c06914fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.942455] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.943207] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c077e02-6858-4182-8bcf-00da2453fa88 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.947926] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 799.947926] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521b78fd-4744-b543-3ab5-e4258182e6b0" [ 799.947926] env[62552]: _type = "Task" [ 799.947926] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.966090] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 714fb65d-9f80-4a81-a637-3e4398405d9b/714fb65d-9f80-4a81-a637-3e4398405d9b.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.969499] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bad3e15d-fa93-441a-bafb-1eae04d03e60 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.989316] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521b78fd-4744-b543-3ab5-e4258182e6b0, 'name': SearchDatastore_Task, 'duration_secs': 0.007705} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.990606] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.990912] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 34736dd0-e617-475e-baa2-cb372db1afb2/34736dd0-e617-475e-baa2-cb372db1afb2.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 799.991294] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 799.991294] env[62552]: value = "task-1239344" [ 799.991294] env[62552]: _type = "Task" [ 799.991294] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.991576] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0401331c-022f-428f-83ec-a72e5ba9fa17 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.002325] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239344, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.002633] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 800.002633] env[62552]: value = "task-1239345" [ 800.002633] env[62552]: _type = "Task" [ 800.002633] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.009964] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239345, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.171978] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 22357d4e-9771-477c-9fc3-fe3d76f6e902 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.283968] env[62552]: DEBUG oslo_vmware.api [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239340, 'name': PowerOnVM_Task, 'duration_secs': 0.516174} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.284271] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 800.284516] env[62552]: INFO nova.compute.manager [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Took 11.50 seconds to spawn the instance on the hypervisor. [ 800.284744] env[62552]: DEBUG nova.compute.manager [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 800.285544] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1690aa6-9f7f-4d3b-b485-519ca165a918 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.320576] env[62552]: DEBUG oslo_concurrency.lockutils [req-fea1e64a-14fe-4aa8-be41-0f2fbfe890ed req-f293fddc-2040-44a1-9515-ba290d52013a service nova] Releasing lock "refresh_cache-34736dd0-e617-475e-baa2-cb372db1afb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.320806] env[62552]: DEBUG nova.compute.manager [req-fea1e64a-14fe-4aa8-be41-0f2fbfe890ed req-f293fddc-2040-44a1-9515-ba290d52013a service nova] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Received event network-vif-deleted-7378d663-07fd-4412-9093-8774fdf7c753 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 800.395721] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239343, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.503767] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239344, 'name': ReconfigVM_Task, 'duration_secs': 0.259799} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.507182] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 714fb65d-9f80-4a81-a637-3e4398405d9b/714fb65d-9f80-4a81-a637-3e4398405d9b.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.507897] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-193e317d-b0c0-435d-a160-eb34c1f5b2f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.514745] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239345, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.516389] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 800.516389] env[62552]: value = "task-1239346" [ 800.516389] env[62552]: _type = "Task" [ 800.516389] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.527020] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239346, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.675200] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance d8891025-5bdf-4dc9-a2b0-c86f94582ac6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.806089] env[62552]: INFO nova.compute.manager [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Took 26.46 seconds to build instance. [ 800.896116] env[62552]: DEBUG oslo_vmware.api [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239343, 'name': PowerOnVM_Task, 'duration_secs': 0.748954} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.896552] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 800.896841] env[62552]: INFO nova.compute.manager [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Took 10.60 seconds to spawn the instance on the hypervisor. [ 800.898056] env[62552]: DEBUG nova.compute.manager [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 800.899529] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b72a750-4017-4f83-a3ef-7264d8bb1277 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.014102] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239345, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.623757} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.014364] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 34736dd0-e617-475e-baa2-cb372db1afb2/34736dd0-e617-475e-baa2-cb372db1afb2.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.014574] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.014817] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-413ada38-faa3-4df0-9846-f1915f5aa2d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.024714] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239346, 'name': Rename_Task, 'duration_secs': 0.154759} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.025998] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 801.026329] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 801.026329] env[62552]: value = "task-1239347" [ 801.026329] env[62552]: _type = "Task" [ 801.026329] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.026513] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c370ca06-a625-404d-a607-1db7cd117278 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.036038] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239347, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.037381] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 801.037381] env[62552]: value = "task-1239348" [ 801.037381] env[62552]: _type = "Task" [ 801.037381] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.044677] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239348, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.110262] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquiring lock "4b80bc41-1b63-444e-8039-696e3d78d9a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.178165] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 4083ee43-ecea-4ea5-8923-42b348893824 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.310405] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b032af0f-d677-4b1b-b982-eebdd4e3f83a tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "4b80bc41-1b63-444e-8039-696e3d78d9a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.133s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.311765] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "4b80bc41-1b63-444e-8039-696e3d78d9a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.202s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.311978] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquiring lock "4b80bc41-1b63-444e-8039-696e3d78d9a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.312199] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "4b80bc41-1b63-444e-8039-696e3d78d9a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.312366] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "4b80bc41-1b63-444e-8039-696e3d78d9a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.318016] env[62552]: INFO nova.compute.manager [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Terminating instance [ 801.424298] env[62552]: INFO nova.compute.manager [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Took 26.98 seconds to build instance. [ 801.538314] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239347, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067533} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.541617] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.542380] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2985bfa8-7e1d-485a-ba1a-d5b3137e0e22 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.549536] env[62552]: DEBUG oslo_vmware.api [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239348, 'name': PowerOnVM_Task, 'duration_secs': 0.498272} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.559496] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.559724] env[62552]: INFO nova.compute.manager [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Took 8.76 seconds to spawn the instance on the hypervisor. [ 801.559903] env[62552]: DEBUG nova.compute.manager [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 801.568679] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 34736dd0-e617-475e-baa2-cb372db1afb2/34736dd0-e617-475e-baa2-cb372db1afb2.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.569413] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da9cf08-9820-43da-a541-0372569800cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.572034] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c074cd67-2863-45d4-8e0e-29f215cb1223 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.594026] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 801.594026] env[62552]: value = "task-1239349" [ 801.594026] env[62552]: _type = "Task" [ 801.594026] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.601917] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239349, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.681932] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance e31e3077-04e1-4adb-a0cc-44cf84a89eda has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.822804] env[62552]: DEBUG nova.compute.manager [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 801.823445] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 801.824139] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 801.827518] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9224db-40be-4673-8fe2-ed9a2f3d50f8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.835259] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 801.835497] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fdeb8f5-aa7e-46bb-bf6c-6d6273c97a0b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.841867] env[62552]: DEBUG oslo_vmware.api [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 801.841867] env[62552]: value = "task-1239350" [ 801.841867] env[62552]: _type = "Task" [ 801.841867] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.849989] env[62552]: DEBUG oslo_vmware.api [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239350, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.927966] env[62552]: DEBUG oslo_concurrency.lockutils [None req-77b71de2-a695-43bf-8c4f-28e4041d6343 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.450s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.104915] env[62552]: INFO nova.compute.manager [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Took 27.61 seconds to build instance. [ 802.109744] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239349, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.185153] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 1cfeedac-f71e-42e4-a04f-8a0462c85907 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.352024] env[62552]: DEBUG oslo_vmware.api [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239350, 'name': PowerOffVM_Task, 'duration_secs': 0.448272} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.352941] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.353219] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 802.353383] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 802.353617] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8b63083-74b7-4e15-a15f-00ac052d92d3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.415759] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 802.415990] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 802.416193] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Deleting the datastore file [datastore1] 4b80bc41-1b63-444e-8039-696e3d78d9a1 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 802.416457] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb0e6a53-bbee-4499-86b3-580e6ab5eb93 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.422736] env[62552]: DEBUG oslo_vmware.api [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for the task: (returnval){ [ 802.422736] env[62552]: value = "task-1239352" [ 802.422736] env[62552]: _type = "Task" [ 802.422736] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.430609] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 802.433197] env[62552]: DEBUG oslo_vmware.api [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239352, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.609798] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c5a0f7b9-5f6a-44b3-91c7-d3b5c6c7ecfb tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "714fb65d-9f80-4a81-a637-3e4398405d9b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.947s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.610090] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239349, 'name': ReconfigVM_Task, 'duration_secs': 0.739127} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.611343] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 34736dd0-e617-475e-baa2-cb372db1afb2/34736dd0-e617-475e-baa2-cb372db1afb2.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.612147] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62124d9c-8a70-4497-bfa6-4f3441667c74 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.617985] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 802.617985] env[62552]: value = "task-1239353" [ 802.617985] env[62552]: _type = "Task" [ 802.617985] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.625772] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239353, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.688971] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance fb881c34-dccc-4703-af8e-c75caafd9b08 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.762752] env[62552]: DEBUG nova.compute.manager [req-3f1bdd35-db84-4257-a7c7-9f7e2dffdf06 req-8c8debc0-1b6f-44e0-adac-d37c87f7091f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Received event network-changed-e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 802.763690] env[62552]: DEBUG nova.compute.manager [req-3f1bdd35-db84-4257-a7c7-9f7e2dffdf06 req-8c8debc0-1b6f-44e0-adac-d37c87f7091f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Refreshing instance network info cache due to event network-changed-e3439c52-55fc-4ea8-ae30-607c154031cb. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 802.764074] env[62552]: DEBUG oslo_concurrency.lockutils [req-3f1bdd35-db84-4257-a7c7-9f7e2dffdf06 req-8c8debc0-1b6f-44e0-adac-d37c87f7091f service nova] Acquiring lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.764276] env[62552]: DEBUG oslo_concurrency.lockutils [req-3f1bdd35-db84-4257-a7c7-9f7e2dffdf06 req-8c8debc0-1b6f-44e0-adac-d37c87f7091f service nova] Acquired lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.764495] env[62552]: DEBUG nova.network.neutron [req-3f1bdd35-db84-4257-a7c7-9f7e2dffdf06 req-8c8debc0-1b6f-44e0-adac-d37c87f7091f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Refreshing network info cache for port e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.932987] env[62552]: DEBUG oslo_vmware.api [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Task: {'id': task-1239352, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135652} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.933311] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 802.933503] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 802.933675] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 802.933846] env[62552]: INFO nova.compute.manager [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 802.934106] env[62552]: DEBUG oslo.service.loopingcall [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.934301] env[62552]: DEBUG nova.compute.manager [-] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 802.934400] env[62552]: DEBUG nova.network.neutron [-] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.956815] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.115433] env[62552]: DEBUG nova.compute.manager [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 803.128010] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239353, 'name': Rename_Task, 'duration_secs': 0.165677} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.128634] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 803.128901] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e770fd50-ed70-4c64-bfbe-75ac085c2f96 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.134662] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 803.134662] env[62552]: value = "task-1239354" [ 803.134662] env[62552]: _type = "Task" [ 803.134662] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.143711] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239354, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.192657] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance fea83dde-3181-49dd-a000-5e5cffc8de95 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.192972] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 803.193134] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 803.245946] env[62552]: INFO nova.compute.manager [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Rescuing [ 803.246069] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "refresh_cache-714fb65d-9f80-4a81-a637-3e4398405d9b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.246228] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "refresh_cache-714fb65d-9f80-4a81-a637-3e4398405d9b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.246395] env[62552]: DEBUG nova.network.neutron [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.613067] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0da81d-cb2a-4f9a-aeef-0fb4a4ed5788 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.623926] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3ab298-c908-48b3-933b-3bd2c7d19568 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.656546] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.660167] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d663120b-301f-4b0d-ba9a-2a90e114769d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.667437] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239354, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.670455] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597d1275-d779-41da-bfd6-ab37cd228191 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.683429] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.755328] env[62552]: DEBUG nova.network.neutron [req-3f1bdd35-db84-4257-a7c7-9f7e2dffdf06 req-8c8debc0-1b6f-44e0-adac-d37c87f7091f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updated VIF entry in instance network info cache for port e3439c52-55fc-4ea8-ae30-607c154031cb. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 803.755743] env[62552]: DEBUG nova.network.neutron [req-3f1bdd35-db84-4257-a7c7-9f7e2dffdf06 req-8c8debc0-1b6f-44e0-adac-d37c87f7091f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updating instance_info_cache with network_info: [{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.942751] env[62552]: DEBUG nova.network.neutron [-] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.971081] env[62552]: DEBUG nova.network.neutron [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Updating instance_info_cache with network_info: [{"id": "71e835ff-cc35-4e84-923d-1931fdb1c527", "address": "fa:16:3e:61:46:82", "network": {"id": "925fdab2-f263-46f3-ab17-2f51407bae3e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1178354148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1716b64900f44a8fb2282801905efd67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e835ff-cc", "ovs_interfaceid": "71e835ff-cc35-4e84-923d-1931fdb1c527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.166285] env[62552]: DEBUG oslo_vmware.api [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239354, 'name': PowerOnVM_Task, 'duration_secs': 0.702365} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.166285] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 804.166285] env[62552]: INFO nova.compute.manager [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Took 8.58 seconds to spawn the instance on the hypervisor. [ 804.166285] env[62552]: DEBUG nova.compute.manager [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 804.167175] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c820de-8a39-47e7-9706-dbbe3e9e8839 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.186898] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 804.261667] env[62552]: DEBUG oslo_concurrency.lockutils [req-3f1bdd35-db84-4257-a7c7-9f7e2dffdf06 req-8c8debc0-1b6f-44e0-adac-d37c87f7091f service nova] Releasing lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.444159] env[62552]: INFO nova.compute.manager [-] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Took 1.51 seconds to deallocate network for instance. [ 804.473960] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "refresh_cache-714fb65d-9f80-4a81-a637-3e4398405d9b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.688125] env[62552]: INFO nova.compute.manager [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Took 26.73 seconds to build instance. [ 804.691323] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 804.691732] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.434s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.691845] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.540s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.693324] env[62552]: INFO nova.compute.claims [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.696912] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 804.696912] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Cleaning up deleted instances {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11571}} [ 804.823404] env[62552]: DEBUG nova.compute.manager [req-86c12408-cd82-4d83-be73-84d6bdefe025 req-549cac08-458c-42fb-b488-86407df5b6ce service nova] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Received event network-vif-deleted-5ca9cef2-ed2e-436e-a641-e24a499e1714 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 804.950578] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.190084] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ee57c86c-6c51-4b9a-a11a-c014ce57e56b tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "34736dd0-e617-475e-baa2-cb372db1afb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.366s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.200069] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] There are 3 instances to clean {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11580}} [ 805.200361] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 9952d8f6-ee78-4c2d-b147-8c08c027f440] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 805.506820] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 805.507165] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3dd9e9f6-9074-4203-acbe-07735583909f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.514126] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 805.514126] env[62552]: value = "task-1239355" [ 805.514126] env[62552]: _type = "Task" [ 805.514126] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.522565] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.693206] env[62552]: DEBUG nova.compute.manager [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 805.705473] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 7408a538-6091-4aa2-b2d0-a3d93840b341] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 805.788837] env[62552]: DEBUG nova.scheduler.client.report [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 805.805238] env[62552]: DEBUG nova.scheduler.client.report [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 805.805556] env[62552]: DEBUG nova.compute.provider_tree [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 805.818186] env[62552]: DEBUG nova.scheduler.client.report [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 805.838226] env[62552]: DEBUG nova.scheduler.client.report [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 806.025823] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239355, 'name': PowerOffVM_Task, 'duration_secs': 0.205264} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.026109] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 806.026955] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a47e7bf-6dc0-4181-add2-ca547fc2cf70 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.049467] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82033395-8600-408f-a4c1-ad0e3ec75f4b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.057105] env[62552]: DEBUG nova.compute.manager [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 806.060317] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9167971e-7f34-463e-b67d-4490eb8c49c1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.080801] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 806.081078] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64ed2207-5144-4e8a-8669-d86c8105ecc8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.087872] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 806.087872] env[62552]: value = "task-1239356" [ 806.087872] env[62552]: _type = "Task" [ 806.087872] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.096480] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239356, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.127915] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e3b1a4-0c3e-4061-9399-59217b66c3fc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.134581] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4744b652-109d-4ebd-b97f-80049d0023e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.163886] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe9979e-88f2-44bc-8b0c-a99b67a6779c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.170180] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfecf7a-ba3c-4574-84e7-ae7e5f00c644 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.182958] env[62552]: DEBUG nova.compute.provider_tree [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.208384] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6240d4e9-79ff-4c84-9d04-49879088bde4] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 806.210647] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.571179] env[62552]: INFO nova.compute.manager [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] instance snapshotting [ 806.574256] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078ba6cd-65c7-495a-b28a-45c58367b245 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.596459] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b940844-4124-4fca-ac58-0a1417faa3f4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.606102] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 806.606329] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.606583] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.606732] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.606912] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.609463] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7c2e563-0673-44cb-9b27-3cd62cc0bcee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.617206] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.617397] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.618102] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c8beed9-f22f-4b3f-b9bb-d74442424aff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.623024] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 806.623024] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524a4bf2-97d2-faf5-e4b1-5be98cb32322" [ 806.623024] env[62552]: _type = "Task" [ 806.623024] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.630166] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524a4bf2-97d2-faf5-e4b1-5be98cb32322, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.685536] env[62552]: DEBUG nova.scheduler.client.report [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 806.711678] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 806.711944] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Cleaning up deleted instances with incomplete migration {{(pid=62552) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11609}} [ 807.112608] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 807.112977] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cc2ef0df-90b3-487a-ad2b-135ba4207d14 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.120311] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 807.120311] env[62552]: value = "task-1239357" [ 807.120311] env[62552]: _type = "Task" [ 807.120311] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.131727] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239357, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.135014] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524a4bf2-97d2-faf5-e4b1-5be98cb32322, 'name': SearchDatastore_Task, 'duration_secs': 0.007568} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.135739] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b64aabc9-512a-4485-a38f-5b92cf4fa0f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.140468] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 807.140468] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dd6a63-af4f-9bd4-da1d-49f45e053d8d" [ 807.140468] env[62552]: _type = "Task" [ 807.140468] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.149214] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dd6a63-af4f-9bd4-da1d-49f45e053d8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.190365] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.190853] env[62552]: DEBUG nova.compute.manager [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 807.193387] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.079s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.194799] env[62552]: INFO nova.compute.claims [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.214715] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.629930] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239357, 'name': CreateSnapshot_Task, 'duration_secs': 0.491921} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.630246] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 807.630982] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc958fcf-d897-4dd8-b0ea-fca0b0146881 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.648889] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dd6a63-af4f-9bd4-da1d-49f45e053d8d, 'name': SearchDatastore_Task, 'duration_secs': 0.009269} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.649150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.649400] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 714fb65d-9f80-4a81-a637-3e4398405d9b/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. {{(pid=62552) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 807.649635] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b009e6d-bed8-4ae5-b35a-6108c5f5e3aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.655643] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 807.655643] env[62552]: value = "task-1239358" [ 807.655643] env[62552]: _type = "Task" [ 807.655643] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.662703] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239358, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.701414] env[62552]: DEBUG nova.compute.utils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.703020] env[62552]: DEBUG nova.compute.manager [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 807.703124] env[62552]: DEBUG nova.network.neutron [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 807.754664] env[62552]: DEBUG nova.policy [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3eaa776d78b04863937ebf10d840b944', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f6b282bb22f464ea5efc77fc17c3004', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 808.087982] env[62552]: DEBUG nova.network.neutron [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Successfully created port: 9de2eca6-381f-431c-9e6f-b422ec21ac62 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.148453] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 808.148768] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-edeed014-d2c2-48ed-b1e1-e591d14df7f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.156699] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 808.156699] env[62552]: value = "task-1239359" [ 808.156699] env[62552]: _type = "Task" [ 808.156699] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.166982] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239358, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446148} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.170234] env[62552]: INFO nova.virt.vmwareapi.ds_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 714fb65d-9f80-4a81-a637-3e4398405d9b/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. [ 808.170538] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239359, 'name': CloneVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.171217] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5955bb54-f60d-4dd1-87e7-e2a02f506703 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.199193] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 714fb65d-9f80-4a81-a637-3e4398405d9b/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.199518] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5fe8f7e-304f-481e-a63b-3e37ee7274e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.216181] env[62552]: DEBUG nova.compute.manager [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 808.223329] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 808.223329] env[62552]: value = "task-1239360" [ 808.223329] env[62552]: _type = "Task" [ 808.223329] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.232560] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239360, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.386502] env[62552]: DEBUG nova.network.neutron [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Successfully created port: fc44bf04-e20d-419e-a8ca-04c24a9fcf2e {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.603952] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7daae8ec-6263-4ba6-ba5d-5c86434a3dde {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.611596] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2acc8741-e099-4ff9-9857-f663796f80b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.649896] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779a27d3-f3d6-453e-992b-afcbee4813d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.657279] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b832d224-faef-4629-9514-525528d85c72 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.673553] env[62552]: DEBUG nova.compute.provider_tree [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.677710] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239359, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.733656] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239360, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.172951] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239359, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.182034] env[62552]: DEBUG nova.scheduler.client.report [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 809.224743] env[62552]: DEBUG nova.compute.manager [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 809.237390] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239360, 'name': ReconfigVM_Task, 'duration_secs': 0.906091} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.238219] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 714fb65d-9f80-4a81-a637-3e4398405d9b/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.239410] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ca4a97-c32c-4a3c-9be3-23cf015060f7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.267018] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2606ea47-c530-4423-9cee-18fa7635646f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.283063] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 809.283063] env[62552]: value = "task-1239361" [ 809.283063] env[62552]: _type = "Task" [ 809.283063] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.284934] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.285254] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.285451] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.285652] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.285805] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.286029] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.286246] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.286411] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.286580] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.286742] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.286913] env[62552]: DEBUG nova.virt.hardware [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.287751] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbceef4-037e-46f5-bb57-4e07723ff4ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.301438] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c5e1d6-56b3-46d4-9dec-c10f360b7c72 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.305905] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.672033] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239359, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.685031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.685522] env[62552]: DEBUG nova.compute.manager [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 809.688068] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.056s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.689425] env[62552]: INFO nova.compute.claims [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.796492] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239361, 'name': ReconfigVM_Task, 'duration_secs': 0.279767} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.796802] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 809.797065] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ecf70e16-d8bd-4faf-b09b-6dec6aed9f44 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.804055] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 809.804055] env[62552]: value = "task-1239362" [ 809.804055] env[62552]: _type = "Task" [ 809.804055] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.811215] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239362, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.046280] env[62552]: DEBUG nova.compute.manager [req-d8feff01-d96d-4e34-9c9a-85f01b19ab51 req-3e58e572-540f-4976-9f17-3bafd4244a7a service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Received event network-vif-plugged-9de2eca6-381f-431c-9e6f-b422ec21ac62 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 810.046475] env[62552]: DEBUG oslo_concurrency.lockutils [req-d8feff01-d96d-4e34-9c9a-85f01b19ab51 req-3e58e572-540f-4976-9f17-3bafd4244a7a service nova] Acquiring lock "36233c83-1170-4a7a-be0b-6a0e8b139de7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.046653] env[62552]: DEBUG oslo_concurrency.lockutils [req-d8feff01-d96d-4e34-9c9a-85f01b19ab51 req-3e58e572-540f-4976-9f17-3bafd4244a7a service nova] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.046835] env[62552]: DEBUG oslo_concurrency.lockutils [req-d8feff01-d96d-4e34-9c9a-85f01b19ab51 req-3e58e572-540f-4976-9f17-3bafd4244a7a service nova] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.047015] env[62552]: DEBUG nova.compute.manager [req-d8feff01-d96d-4e34-9c9a-85f01b19ab51 req-3e58e572-540f-4976-9f17-3bafd4244a7a service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] No waiting events found dispatching network-vif-plugged-9de2eca6-381f-431c-9e6f-b422ec21ac62 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 810.047263] env[62552]: WARNING nova.compute.manager [req-d8feff01-d96d-4e34-9c9a-85f01b19ab51 req-3e58e572-540f-4976-9f17-3bafd4244a7a service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Received unexpected event network-vif-plugged-9de2eca6-381f-431c-9e6f-b422ec21ac62 for instance with vm_state building and task_state spawning. [ 810.130723] env[62552]: DEBUG nova.network.neutron [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Successfully updated port: 9de2eca6-381f-431c-9e6f-b422ec21ac62 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 810.174612] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239359, 'name': CloneVM_Task} progress is 95%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.193977] env[62552]: DEBUG nova.compute.utils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.197903] env[62552]: DEBUG nova.compute.manager [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 810.198069] env[62552]: DEBUG nova.network.neutron [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.239952] env[62552]: DEBUG nova.policy [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7f8aa8200874dddb71d8b21bd12ca04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8b8d96b464a439e9c7ef6f3e419a9bc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.314076] env[62552]: DEBUG oslo_vmware.api [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239362, 'name': PowerOnVM_Task, 'duration_secs': 0.399696} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.314347] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 810.316978] env[62552]: DEBUG nova.compute.manager [None req-37e95e80-9f5d-45f4-9221-f7d2f8f970d9 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 810.317804] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f26262-0160-4104-8198-2d14200150e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.668915] env[62552]: DEBUG nova.network.neutron [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Successfully created port: 3d60a1e0-8a29-42e5-a18e-f2a265bdb330 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.676248] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239359, 'name': CloneVM_Task, 'duration_secs': 2.475589} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.676916] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Created linked-clone VM from snapshot [ 810.677721] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e43e6e-87cc-4d85-9ee9-d74c0041c7a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.687407] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Uploading image 35ac675b-d35f-462e-aaa2-a2061f578780 {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 810.701210] env[62552]: DEBUG nova.compute.manager [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 810.720097] env[62552]: DEBUG oslo_vmware.rw_handles [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 810.720097] env[62552]: value = "vm-267390" [ 810.720097] env[62552]: _type = "VirtualMachine" [ 810.720097] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 810.720097] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-201d593b-f52e-4cb1-8c76-75a744b1a1ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.731606] env[62552]: DEBUG oslo_vmware.rw_handles [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lease: (returnval){ [ 810.731606] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52422a0d-802d-2c3d-b53a-8a83bc43c07b" [ 810.731606] env[62552]: _type = "HttpNfcLease" [ 810.731606] env[62552]: } obtained for exporting VM: (result){ [ 810.731606] env[62552]: value = "vm-267390" [ 810.731606] env[62552]: _type = "VirtualMachine" [ 810.731606] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 810.731606] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the lease: (returnval){ [ 810.731606] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52422a0d-802d-2c3d-b53a-8a83bc43c07b" [ 810.731606] env[62552]: _type = "HttpNfcLease" [ 810.731606] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 810.737400] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 810.737400] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52422a0d-802d-2c3d-b53a-8a83bc43c07b" [ 810.737400] env[62552]: _type = "HttpNfcLease" [ 810.737400] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 811.081232] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cac437-ad1b-4ab9-8b76-3ff5a530b7e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.088311] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d495c59-5e32-4f4c-b029-0affc117abe8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.121479] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7bdfaef-789a-4f7e-94c5-6b04f1835de2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.133819] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84915f68-443d-455d-86fa-18979526096c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.152073] env[62552]: DEBUG nova.compute.provider_tree [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.238693] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 811.238693] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52422a0d-802d-2c3d-b53a-8a83bc43c07b" [ 811.238693] env[62552]: _type = "HttpNfcLease" [ 811.238693] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 811.239045] env[62552]: DEBUG oslo_vmware.rw_handles [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 811.239045] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52422a0d-802d-2c3d-b53a-8a83bc43c07b" [ 811.239045] env[62552]: _type = "HttpNfcLease" [ 811.239045] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 811.239770] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd4da60-cb55-451d-bda5-fe959ef19410 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.248023] env[62552]: DEBUG oslo_vmware.rw_handles [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221cf72-750d-6c1e-7a7d-f40405ac2262/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 811.248023] env[62552]: DEBUG oslo_vmware.rw_handles [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221cf72-750d-6c1e-7a7d-f40405ac2262/disk-0.vmdk for reading. {{(pid=62552) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 811.348129] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9c269542-c002-41df-a892-9564c85f7ec2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.655793] env[62552]: DEBUG nova.scheduler.client.report [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 811.719348] env[62552]: DEBUG nova.compute.manager [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 811.754405] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.754672] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.754829] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.755014] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.755177] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.755409] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.755530] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.755688] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.755955] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.756167] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.756347] env[62552]: DEBUG nova.virt.hardware [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.757275] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f34efa-80f7-47fa-9e42-993a75ef2522 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.767788] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ff4b03-2bb4-4183-af7b-45825e408f91 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.163611] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.164428] env[62552]: DEBUG nova.compute.manager [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 812.167792] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.922s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.169798] env[62552]: INFO nova.compute.claims [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.197910] env[62552]: DEBUG nova.compute.manager [req-88ec3fd8-00b0-4f9d-ae79-8f605c5dd990 req-f36243c4-59bc-4ec5-bf5e-d81b1667c482 service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Received event network-changed-9de2eca6-381f-431c-9e6f-b422ec21ac62 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 812.197910] env[62552]: DEBUG nova.compute.manager [req-88ec3fd8-00b0-4f9d-ae79-8f605c5dd990 req-f36243c4-59bc-4ec5-bf5e-d81b1667c482 service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Refreshing instance network info cache due to event network-changed-9de2eca6-381f-431c-9e6f-b422ec21ac62. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 812.198546] env[62552]: DEBUG oslo_concurrency.lockutils [req-88ec3fd8-00b0-4f9d-ae79-8f605c5dd990 req-f36243c4-59bc-4ec5-bf5e-d81b1667c482 service nova] Acquiring lock "refresh_cache-36233c83-1170-4a7a-be0b-6a0e8b139de7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.198856] env[62552]: DEBUG oslo_concurrency.lockutils [req-88ec3fd8-00b0-4f9d-ae79-8f605c5dd990 req-f36243c4-59bc-4ec5-bf5e-d81b1667c482 service nova] Acquired lock "refresh_cache-36233c83-1170-4a7a-be0b-6a0e8b139de7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.199070] env[62552]: DEBUG nova.network.neutron [req-88ec3fd8-00b0-4f9d-ae79-8f605c5dd990 req-f36243c4-59bc-4ec5-bf5e-d81b1667c482 service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Refreshing network info cache for port 9de2eca6-381f-431c-9e6f-b422ec21ac62 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.285644] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "5cee4242-49ae-4cb7-a208-e2982f52fbad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.286711] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "5cee4242-49ae-4cb7-a208-e2982f52fbad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.675703] env[62552]: DEBUG nova.compute.utils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.680814] env[62552]: DEBUG nova.compute.manager [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 812.680814] env[62552]: DEBUG nova.network.neutron [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.807790] env[62552]: DEBUG nova.network.neutron [req-88ec3fd8-00b0-4f9d-ae79-8f605c5dd990 req-f36243c4-59bc-4ec5-bf5e-d81b1667c482 service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.906649] env[62552]: DEBUG nova.policy [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd61cdc9f43084fdda0cc0dabfc76c58e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2f9aa2ab2de4effbcbc7ea69c6db7dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.915128] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "cdee22a9-4327-47af-ab14-8403a06b2802" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.915534] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.915767] env[62552]: INFO nova.compute.manager [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Shelving [ 813.072671] env[62552]: DEBUG nova.network.neutron [req-88ec3fd8-00b0-4f9d-ae79-8f605c5dd990 req-f36243c4-59bc-4ec5-bf5e-d81b1667c482 service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.090667] env[62552]: DEBUG nova.network.neutron [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Successfully updated port: 3d60a1e0-8a29-42e5-a18e-f2a265bdb330 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 813.138281] env[62552]: DEBUG nova.network.neutron [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Successfully updated port: fc44bf04-e20d-419e-a8ca-04c24a9fcf2e {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 813.182061] env[62552]: DEBUG nova.compute.manager [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 813.191013] env[62552]: DEBUG nova.network.neutron [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Successfully created port: 8cf42b44-1622-449e-8a17-e2be57435ee8 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.578612] env[62552]: DEBUG oslo_concurrency.lockutils [req-88ec3fd8-00b0-4f9d-ae79-8f605c5dd990 req-f36243c4-59bc-4ec5-bf5e-d81b1667c482 service nova] Releasing lock "refresh_cache-36233c83-1170-4a7a-be0b-6a0e8b139de7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.595547] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-3b333977-0b4b-4ccf-bed5-231215abae3c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.595547] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-3b333977-0b4b-4ccf-bed5-231215abae3c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.595547] env[62552]: DEBUG nova.network.neutron [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.629889] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49b6721-cffe-472e-9fb4-cc86debd0fb7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.638671] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b28faa-e246-440e-9af6-0f0570da9724 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.642204] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "refresh_cache-36233c83-1170-4a7a-be0b-6a0e8b139de7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.642392] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquired lock "refresh_cache-36233c83-1170-4a7a-be0b-6a0e8b139de7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.642527] env[62552]: DEBUG nova.network.neutron [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.672482] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d41c6d-33f7-4a46-a629-e0daf9896903 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.680682] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53567af5-bc44-41ce-bbb8-114f8efaae4e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.696071] env[62552]: DEBUG nova.compute.provider_tree [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.925782] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 813.926204] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86e0042f-3557-4770-9844-c732cdba9283 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.934818] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 813.934818] env[62552]: value = "task-1239364" [ 813.934818] env[62552]: _type = "Task" [ 813.934818] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.945298] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.149995] env[62552]: DEBUG nova.network.neutron [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.199328] env[62552]: DEBUG nova.compute.manager [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 814.201996] env[62552]: DEBUG nova.scheduler.client.report [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 814.228113] env[62552]: DEBUG nova.network.neutron [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.232963] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.233441] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.233561] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.233778] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.233976] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.234194] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.234445] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.234656] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.234881] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.235094] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.235313] env[62552]: DEBUG nova.virt.hardware [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.237045] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1677fd10-d0f2-4a40-aa34-9139ea9e4ca7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.245535] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe21f9b0-2883-47a2-a21c-b5e8d5672a7b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.256837] env[62552]: DEBUG nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Received event network-vif-plugged-3d60a1e0-8a29-42e5-a18e-f2a265bdb330 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 814.257083] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Acquiring lock "3b333977-0b4b-4ccf-bed5-231215abae3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.257306] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.257550] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.257638] env[62552]: DEBUG nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] No waiting events found dispatching network-vif-plugged-3d60a1e0-8a29-42e5-a18e-f2a265bdb330 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 814.257803] env[62552]: WARNING nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Received unexpected event network-vif-plugged-3d60a1e0-8a29-42e5-a18e-f2a265bdb330 for instance with vm_state building and task_state spawning. [ 814.257971] env[62552]: DEBUG nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Received event network-vif-plugged-fc44bf04-e20d-419e-a8ca-04c24a9fcf2e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 814.258174] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Acquiring lock "36233c83-1170-4a7a-be0b-6a0e8b139de7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.258399] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.258557] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.258720] env[62552]: DEBUG nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] No waiting events found dispatching network-vif-plugged-fc44bf04-e20d-419e-a8ca-04c24a9fcf2e {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 814.258882] env[62552]: WARNING nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Received unexpected event network-vif-plugged-fc44bf04-e20d-419e-a8ca-04c24a9fcf2e for instance with vm_state building and task_state spawning. [ 814.259063] env[62552]: DEBUG nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Received event network-changed-3d60a1e0-8a29-42e5-a18e-f2a265bdb330 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 814.259204] env[62552]: DEBUG nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Refreshing instance network info cache due to event network-changed-3d60a1e0-8a29-42e5-a18e-f2a265bdb330. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 814.259496] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Acquiring lock "refresh_cache-3b333977-0b4b-4ccf-bed5-231215abae3c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.444620] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239364, 'name': PowerOffVM_Task, 'duration_secs': 0.219397} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.444917] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 814.445736] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e26743-f69d-4e6a-985a-807a33482e3e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.468580] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544a4ea6-f3f3-4e03-88c9-b22566aae294 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.514673] env[62552]: DEBUG nova.network.neutron [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Updating instance_info_cache with network_info: [{"id": "3d60a1e0-8a29-42e5-a18e-f2a265bdb330", "address": "fa:16:3e:4e:b5:bc", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d60a1e0-8a", "ovs_interfaceid": "3d60a1e0-8a29-42e5-a18e-f2a265bdb330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.711392] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.711967] env[62552]: DEBUG nova.compute.manager [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 814.714977] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.385s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.716959] env[62552]: INFO nova.compute.claims [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.748109] env[62552]: DEBUG nova.network.neutron [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Successfully updated port: 8cf42b44-1622-449e-8a17-e2be57435ee8 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.873728] env[62552]: DEBUG nova.compute.manager [req-a503af8c-f600-4da3-af03-2f099b0e6bde req-aa9e0485-7513-4e66-9fff-6ab089bc10c0 service nova] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Received event network-vif-plugged-8cf42b44-1622-449e-8a17-e2be57435ee8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 814.873965] env[62552]: DEBUG oslo_concurrency.lockutils [req-a503af8c-f600-4da3-af03-2f099b0e6bde req-aa9e0485-7513-4e66-9fff-6ab089bc10c0 service nova] Acquiring lock "fef94e30-f946-4d7f-a108-f32cad505bc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.874398] env[62552]: DEBUG oslo_concurrency.lockutils [req-a503af8c-f600-4da3-af03-2f099b0e6bde req-aa9e0485-7513-4e66-9fff-6ab089bc10c0 service nova] Lock "fef94e30-f946-4d7f-a108-f32cad505bc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.874398] env[62552]: DEBUG oslo_concurrency.lockutils [req-a503af8c-f600-4da3-af03-2f099b0e6bde req-aa9e0485-7513-4e66-9fff-6ab089bc10c0 service nova] Lock "fef94e30-f946-4d7f-a108-f32cad505bc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.874743] env[62552]: DEBUG nova.compute.manager [req-a503af8c-f600-4da3-af03-2f099b0e6bde req-aa9e0485-7513-4e66-9fff-6ab089bc10c0 service nova] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] No waiting events found dispatching network-vif-plugged-8cf42b44-1622-449e-8a17-e2be57435ee8 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 814.874743] env[62552]: WARNING nova.compute.manager [req-a503af8c-f600-4da3-af03-2f099b0e6bde req-aa9e0485-7513-4e66-9fff-6ab089bc10c0 service nova] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Received unexpected event network-vif-plugged-8cf42b44-1622-449e-8a17-e2be57435ee8 for instance with vm_state building and task_state spawning. [ 814.959973] env[62552]: DEBUG nova.network.neutron [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Updating instance_info_cache with network_info: [{"id": "9de2eca6-381f-431c-9e6f-b422ec21ac62", "address": "fa:16:3e:ae:ca:2c", "network": {"id": "957dd04b-7da7-4e7c-99ad-f15d09c64588", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-444708671", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.250", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f6b282bb22f464ea5efc77fc17c3004", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9de2eca6-38", "ovs_interfaceid": "9de2eca6-381f-431c-9e6f-b422ec21ac62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fc44bf04-e20d-419e-a8ca-04c24a9fcf2e", "address": "fa:16:3e:04:76:0b", "network": {"id": "cf1c2c04-c578-4e1c-acc4-c2787966a28c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1425449914", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.194", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "0f6b282bb22f464ea5efc77fc17c3004", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc44bf04-e2", "ovs_interfaceid": "fc44bf04-e20d-419e-a8ca-04c24a9fcf2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.980492] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 814.981487] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9df3cd9c-491e-4570-b21d-caffb0eba43a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.989255] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 814.989255] env[62552]: value = "task-1239365" [ 814.989255] env[62552]: _type = "Task" [ 814.989255] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.999120] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239365, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.016698] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-3b333977-0b4b-4ccf-bed5-231215abae3c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.017077] env[62552]: DEBUG nova.compute.manager [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Instance network_info: |[{"id": "3d60a1e0-8a29-42e5-a18e-f2a265bdb330", "address": "fa:16:3e:4e:b5:bc", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d60a1e0-8a", "ovs_interfaceid": "3d60a1e0-8a29-42e5-a18e-f2a265bdb330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 815.017435] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Acquired lock "refresh_cache-3b333977-0b4b-4ccf-bed5-231215abae3c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.017723] env[62552]: DEBUG nova.network.neutron [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Refreshing network info cache for port 3d60a1e0-8a29-42e5-a18e-f2a265bdb330 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 815.018928] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:b5:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d60a1e0-8a29-42e5-a18e-f2a265bdb330', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.026623] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Creating folder: Project (b8b8d96b464a439e9c7ef6f3e419a9bc). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.027226] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c68a8f4d-9b54-43b9-b569-709d4f7e82a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.038395] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Created folder: Project (b8b8d96b464a439e9c7ef6f3e419a9bc) in parent group-v267339. [ 815.038696] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Creating folder: Instances. Parent ref: group-v267391. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.039041] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e6ca4a61-44fc-4ee5-9d8b-4f2f92f82057 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.048084] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Created folder: Instances in parent group-v267391. [ 815.048373] env[62552]: DEBUG oslo.service.loopingcall [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.048591] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 815.048817] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e72ca6e-632d-4a4a-b9e4-17d6dba960d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.068795] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.068795] env[62552]: value = "task-1239368" [ 815.068795] env[62552]: _type = "Task" [ 815.068795] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.076461] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239368, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.224158] env[62552]: DEBUG nova.compute.utils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.229022] env[62552]: DEBUG nova.compute.manager [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 815.229022] env[62552]: DEBUG nova.network.neutron [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 815.250869] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquiring lock "refresh_cache-fef94e30-f946-4d7f-a108-f32cad505bc4" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.250978] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquired lock "refresh_cache-fef94e30-f946-4d7f-a108-f32cad505bc4" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.252023] env[62552]: DEBUG nova.network.neutron [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.286099] env[62552]: DEBUG nova.policy [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3faa654bd214bbb886d692e9ea29bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'adb0b1256dd34e33913d5ae5bb3553b5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.464103] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Releasing lock "refresh_cache-36233c83-1170-4a7a-be0b-6a0e8b139de7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.464644] env[62552]: DEBUG nova.compute.manager [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Instance network_info: |[{"id": "9de2eca6-381f-431c-9e6f-b422ec21ac62", "address": "fa:16:3e:ae:ca:2c", "network": {"id": "957dd04b-7da7-4e7c-99ad-f15d09c64588", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-444708671", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.250", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f6b282bb22f464ea5efc77fc17c3004", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9de2eca6-38", "ovs_interfaceid": "9de2eca6-381f-431c-9e6f-b422ec21ac62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fc44bf04-e20d-419e-a8ca-04c24a9fcf2e", "address": "fa:16:3e:04:76:0b", "network": {"id": "cf1c2c04-c578-4e1c-acc4-c2787966a28c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1425449914", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.194", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "0f6b282bb22f464ea5efc77fc17c3004", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc44bf04-e2", "ovs_interfaceid": "fc44bf04-e20d-419e-a8ca-04c24a9fcf2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 815.465643] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:ca:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4576b9d4-535c-40aa-b078-246f671f216e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9de2eca6-381f-431c-9e6f-b422ec21ac62', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:76:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fc44bf04-e20d-419e-a8ca-04c24a9fcf2e', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.475990] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Creating folder: Project (0f6b282bb22f464ea5efc77fc17c3004). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.476736] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28be29d1-8daa-414a-b45f-dc426a264fcd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.486678] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Created folder: Project (0f6b282bb22f464ea5efc77fc17c3004) in parent group-v267339. [ 815.486967] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Creating folder: Instances. Parent ref: group-v267394. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.487281] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30123225-1aff-4f47-900a-fa8d44232aab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.498537] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Created folder: Instances in parent group-v267394. [ 815.498537] env[62552]: DEBUG oslo.service.loopingcall [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.498537] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 815.498753] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b250ecb-ff26-463b-9b48-3626a4af6b70 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.520430] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239365, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.524533] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.524533] env[62552]: value = "task-1239371" [ 815.524533] env[62552]: _type = "Task" [ 815.524533] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.535246] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239371, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.580837] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239368, 'name': CreateVM_Task, 'duration_secs': 0.429943} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.580837] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 815.580837] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.580837] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.580837] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.581134] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c17407c0-54f2-478b-aebd-b516e1d40d6c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.586120] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 815.586120] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f7eb60-0eb7-125b-41ee-435b235e85d0" [ 815.586120] env[62552]: _type = "Task" [ 815.586120] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.594998] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f7eb60-0eb7-125b-41ee-435b235e85d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.732197] env[62552]: DEBUG nova.compute.manager [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 815.818820] env[62552]: DEBUG nova.network.neutron [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Updated VIF entry in instance network info cache for port 3d60a1e0-8a29-42e5-a18e-f2a265bdb330. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 815.819093] env[62552]: DEBUG nova.network.neutron [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Updating instance_info_cache with network_info: [{"id": "3d60a1e0-8a29-42e5-a18e-f2a265bdb330", "address": "fa:16:3e:4e:b5:bc", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d60a1e0-8a", "ovs_interfaceid": "3d60a1e0-8a29-42e5-a18e-f2a265bdb330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.838053] env[62552]: DEBUG nova.network.neutron [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.843912] env[62552]: DEBUG nova.network.neutron [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Successfully created port: 76bce90f-8d24-4e1e-8562-f2790b183627 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.934021] env[62552]: INFO nova.compute.manager [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Rebuilding instance [ 815.988488] env[62552]: DEBUG nova.compute.manager [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 815.989593] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b06fc8-2608-4f86-a028-12fd76ac6952 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.013724] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239365, 'name': CreateSnapshot_Task, 'duration_secs': 0.722681} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.014515] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 816.016335] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a50358-d4e0-4df0-9031-813e526e797f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.039372] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239371, 'name': CreateVM_Task, 'duration_secs': 0.426768} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.039705] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.041036] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.097181] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f7eb60-0eb7-125b-41ee-435b235e85d0, 'name': SearchDatastore_Task, 'duration_secs': 0.013382} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.100286] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.100738] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.101139] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.104026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.104026] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.104026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.104026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.104026] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4eb68a28-0cc6-42f5-b4d8-8b179586c3b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.105187] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-133a4170-38c1-4d8d-97dd-4c05d2e0c39e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.110138] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 816.110138] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5243ba6c-2171-62c5-9682-71cb8357d550" [ 816.110138] env[62552]: _type = "Task" [ 816.110138] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.119025] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.119423] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.121014] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2cd7cd4-e7ce-4e3f-99fe-2ec41a2d753b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.126453] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5243ba6c-2171-62c5-9682-71cb8357d550, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.129432] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 816.129432] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bdef72-60dc-132d-ab6e-5b96cb506d89" [ 816.129432] env[62552]: _type = "Task" [ 816.129432] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.144021] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bdef72-60dc-132d-ab6e-5b96cb506d89, 'name': SearchDatastore_Task, 'duration_secs': 0.007863} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.144021] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8be7834c-0207-4349-8483-490a97d11395 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.153137] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 816.153137] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a2b114-2e46-9d11-e2d5-b5cbe1d62810" [ 816.153137] env[62552]: _type = "Task" [ 816.153137] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.159759] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a2b114-2e46-9d11-e2d5-b5cbe1d62810, 'name': SearchDatastore_Task, 'duration_secs': 0.007644} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.160231] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.160891] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 3b333977-0b4b-4ccf-bed5-231215abae3c/3b333977-0b4b-4ccf-bed5-231215abae3c.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 816.163719] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0bc7eb0e-9a42-4f2e-8d45-54bb3a07194c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.173344] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 816.173344] env[62552]: value = "task-1239372" [ 816.173344] env[62552]: _type = "Task" [ 816.173344] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.186443] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239372, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.213321] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110dfa90-248d-40d5-8df7-b7145ac86625 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.220517] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2f93c9-f837-496a-9cbb-07b9f5bfacb9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.260562] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566507f6-1dab-4959-9f96-8ada5a50a809 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.269148] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56016aa9-5a82-43e7-86c0-cb48cedf7744 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.284290] env[62552]: DEBUG nova.compute.provider_tree [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.294869] env[62552]: DEBUG nova.network.neutron [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Updating instance_info_cache with network_info: [{"id": "8cf42b44-1622-449e-8a17-e2be57435ee8", "address": "fa:16:3e:c3:f5:7e", "network": {"id": "f5ce996d-7e56-4f9f-a7c2-f1adbc556126", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-601296890-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f9aa2ab2de4effbcbc7ea69c6db7dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cf42b44-16", "ovs_interfaceid": "8cf42b44-1622-449e-8a17-e2be57435ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.324618] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Releasing lock "refresh_cache-3b333977-0b4b-4ccf-bed5-231215abae3c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.324618] env[62552]: DEBUG nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Received event network-changed-fc44bf04-e20d-419e-a8ca-04c24a9fcf2e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 816.324799] env[62552]: DEBUG nova.compute.manager [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Refreshing instance network info cache due to event network-changed-fc44bf04-e20d-419e-a8ca-04c24a9fcf2e. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 816.325212] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Acquiring lock "refresh_cache-36233c83-1170-4a7a-be0b-6a0e8b139de7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.325519] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Acquired lock "refresh_cache-36233c83-1170-4a7a-be0b-6a0e8b139de7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.325850] env[62552]: DEBUG nova.network.neutron [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Refreshing network info cache for port fc44bf04-e20d-419e-a8ca-04c24a9fcf2e {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.535834] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 816.536635] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1c4c2942-7dc0-4c7b-8e4b-8e7f735a916e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.548509] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 816.548509] env[62552]: value = "task-1239373" [ 816.548509] env[62552]: _type = "Task" [ 816.548509] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.560488] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239373, 'name': CloneVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.623491] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5243ba6c-2171-62c5-9682-71cb8357d550, 'name': SearchDatastore_Task, 'duration_secs': 0.014979} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.623901] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.624170] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.624424] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.624579] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.625247] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.625247] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa100c0f-d06e-404d-9ca7-fded72f51c66 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.645744] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.645744] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.646444] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66fb3b49-2d5b-4c2b-86c8-66c69dc40ceb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.652136] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 816.652136] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525f5306-25e3-c3a1-eeda-b559d126a7de" [ 816.652136] env[62552]: _type = "Task" [ 816.652136] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.661726] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525f5306-25e3-c3a1-eeda-b559d126a7de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.681593] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239372, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486526} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.681897] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 3b333977-0b4b-4ccf-bed5-231215abae3c/3b333977-0b4b-4ccf-bed5-231215abae3c.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 816.682152] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.682428] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-13f58289-b7c4-4f48-b193-38826024298f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.688582] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 816.688582] env[62552]: value = "task-1239374" [ 816.688582] env[62552]: _type = "Task" [ 816.688582] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.697062] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239374, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.761590] env[62552]: DEBUG nova.compute.manager [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 816.785709] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.785978] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.786177] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.786369] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.786523] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.786677] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.786893] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.787384] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.787684] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.787926] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.788196] env[62552]: DEBUG nova.virt.hardware [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.789211] env[62552]: DEBUG nova.scheduler.client.report [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 816.793724] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48bcf6c-a6b1-4a13-ad70-5a1ab81cffc9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.798092] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Releasing lock "refresh_cache-fef94e30-f946-4d7f-a108-f32cad505bc4" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.800151] env[62552]: DEBUG nova.compute.manager [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Instance network_info: |[{"id": "8cf42b44-1622-449e-8a17-e2be57435ee8", "address": "fa:16:3e:c3:f5:7e", "network": {"id": "f5ce996d-7e56-4f9f-a7c2-f1adbc556126", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-601296890-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f9aa2ab2de4effbcbc7ea69c6db7dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cf42b44-16", "ovs_interfaceid": "8cf42b44-1622-449e-8a17-e2be57435ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 816.800151] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:f5:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd998416-f3d6-4a62-b828-5011063ce76a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8cf42b44-1622-449e-8a17-e2be57435ee8', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.807218] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Creating folder: Project (f2f9aa2ab2de4effbcbc7ea69c6db7dd). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.810152] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74965d0b-2b6b-4e47-9d2d-66c9bda64828 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.817037] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc165d4-0c06-448b-898d-3d08fa788be0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.822767] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Created folder: Project (f2f9aa2ab2de4effbcbc7ea69c6db7dd) in parent group-v267339. [ 816.823088] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Creating folder: Instances. Parent ref: group-v267399. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.823671] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-446fea4f-a210-430e-b696-fbc98916a5da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.847105] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Created folder: Instances in parent group-v267399. [ 816.847410] env[62552]: DEBUG oslo.service.loopingcall [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.847588] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 816.847817] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c639d241-e792-41df-8cf5-6bc052c33d4f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.866965] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.866965] env[62552]: value = "task-1239377" [ 816.866965] env[62552]: _type = "Task" [ 816.866965] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.875477] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239377, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.941172] env[62552]: DEBUG nova.compute.manager [req-401809a2-ad32-4970-8b09-9e925f82f9cf req-3951c57a-acfe-4a96-a185-69bfa8a173c1 service nova] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Received event network-changed-8cf42b44-1622-449e-8a17-e2be57435ee8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 816.941409] env[62552]: DEBUG nova.compute.manager [req-401809a2-ad32-4970-8b09-9e925f82f9cf req-3951c57a-acfe-4a96-a185-69bfa8a173c1 service nova] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Refreshing instance network info cache due to event network-changed-8cf42b44-1622-449e-8a17-e2be57435ee8. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 816.941687] env[62552]: DEBUG oslo_concurrency.lockutils [req-401809a2-ad32-4970-8b09-9e925f82f9cf req-3951c57a-acfe-4a96-a185-69bfa8a173c1 service nova] Acquiring lock "refresh_cache-fef94e30-f946-4d7f-a108-f32cad505bc4" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.942037] env[62552]: DEBUG oslo_concurrency.lockutils [req-401809a2-ad32-4970-8b09-9e925f82f9cf req-3951c57a-acfe-4a96-a185-69bfa8a173c1 service nova] Acquired lock "refresh_cache-fef94e30-f946-4d7f-a108-f32cad505bc4" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.942110] env[62552]: DEBUG nova.network.neutron [req-401809a2-ad32-4970-8b09-9e925f82f9cf req-3951c57a-acfe-4a96-a185-69bfa8a173c1 service nova] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Refreshing network info cache for port 8cf42b44-1622-449e-8a17-e2be57435ee8 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 817.012777] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 817.013143] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b7bd464-6743-41b2-97e3-3e14d1a7b873 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.021116] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 817.021116] env[62552]: value = "task-1239378" [ 817.021116] env[62552]: _type = "Task" [ 817.021116] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.033226] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239378, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.058627] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239373, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.163035] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525f5306-25e3-c3a1-eeda-b559d126a7de, 'name': SearchDatastore_Task, 'duration_secs': 0.012122} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.163903] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fca0cf39-8c68-41e9-b859-74336bb74cef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.170265] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 817.170265] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526a9872-65c5-0ef3-07bd-55070fdde09d" [ 817.170265] env[62552]: _type = "Task" [ 817.170265] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.178421] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526a9872-65c5-0ef3-07bd-55070fdde09d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.191070] env[62552]: DEBUG nova.network.neutron [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Updated VIF entry in instance network info cache for port fc44bf04-e20d-419e-a8ca-04c24a9fcf2e. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 817.191070] env[62552]: DEBUG nova.network.neutron [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Updating instance_info_cache with network_info: [{"id": "9de2eca6-381f-431c-9e6f-b422ec21ac62", "address": "fa:16:3e:ae:ca:2c", "network": {"id": "957dd04b-7da7-4e7c-99ad-f15d09c64588", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-444708671", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.250", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f6b282bb22f464ea5efc77fc17c3004", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9de2eca6-38", "ovs_interfaceid": "9de2eca6-381f-431c-9e6f-b422ec21ac62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fc44bf04-e20d-419e-a8ca-04c24a9fcf2e", "address": "fa:16:3e:04:76:0b", "network": {"id": "cf1c2c04-c578-4e1c-acc4-c2787966a28c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1425449914", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.194", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "0f6b282bb22f464ea5efc77fc17c3004", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc44bf04-e2", "ovs_interfaceid": "fc44bf04-e20d-419e-a8ca-04c24a9fcf2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.202350] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239374, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111828} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.203587] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.204704] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961b261c-25ff-439e-bb4b-8436364132e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.229506] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 3b333977-0b4b-4ccf-bed5-231215abae3c/3b333977-0b4b-4ccf-bed5-231215abae3c.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.230426] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8c17e7f-baed-426f-9eae-411d7e73cc35 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.251754] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 817.251754] env[62552]: value = "task-1239379" [ 817.251754] env[62552]: _type = "Task" [ 817.251754] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.260582] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239379, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.299024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.584s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.299580] env[62552]: DEBUG nova.compute.manager [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 817.302657] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.942s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.302898] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.306274] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.471s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.308062] env[62552]: INFO nova.compute.claims [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.380964] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239377, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.423325] env[62552]: INFO nova.scheduler.client.report [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Deleted allocations for instance 1472ca01-ba04-4e7d-a9cf-ccaf29c04282 [ 817.537363] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239378, 'name': PowerOffVM_Task, 'duration_secs': 0.258264} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.537596] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 817.538391] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 817.538665] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d37d971f-e224-4dc4-b46e-0717d9e7c81f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.546336] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 817.546336] env[62552]: value = "task-1239380" [ 817.546336] env[62552]: _type = "Task" [ 817.546336] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.562816] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239373, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.563207] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 817.563484] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 817.565310] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267350', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'name': 'volume-0feb401f-606c-4b63-8c60-46d9d717225f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8707ce4b-677e-4f13-86f8-3e327d19380b', 'attached_at': '', 'detached_at': '', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'serial': '0feb401f-606c-4b63-8c60-46d9d717225f'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 817.565310] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff24130-3ccb-476a-8857-1f57ec3a6c74 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.586278] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73df664c-532a-4662-a782-a1d35cbddc10 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.596843] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139ba860-eae0-4449-90fc-bebcec5f1cd1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.616820] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69886de1-66f4-4f28-a5c5-7e48e9ebe179 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.633146] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] The volume has not been displaced from its original location: [datastore2] volume-0feb401f-606c-4b63-8c60-46d9d717225f/volume-0feb401f-606c-4b63-8c60-46d9d717225f.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 817.641142] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Reconfiguring VM instance instance-0000002f to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 817.641253] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fdda671d-20bf-4db8-8567-b2d17d1c5e96 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.660257] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 817.660257] env[62552]: value = "task-1239381" [ 817.660257] env[62552]: _type = "Task" [ 817.660257] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.672029] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239381, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.681116] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526a9872-65c5-0ef3-07bd-55070fdde09d, 'name': SearchDatastore_Task, 'duration_secs': 0.019425} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.681467] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.681759] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 36233c83-1170-4a7a-be0b-6a0e8b139de7/36233c83-1170-4a7a-be0b-6a0e8b139de7.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.682063] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e0fd0a40-32a3-4215-bb3a-8624d8969cfe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.688349] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 817.688349] env[62552]: value = "task-1239382" [ 817.688349] env[62552]: _type = "Task" [ 817.688349] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.693658] env[62552]: DEBUG oslo_concurrency.lockutils [req-a583cfcb-a63f-4242-8807-f1788eb1d535 req-c0543a0a-7d7b-42c0-ad5a-33803ee6138f service nova] Releasing lock "refresh_cache-36233c83-1170-4a7a-be0b-6a0e8b139de7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.697955] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239382, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.769854] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239379, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.813576] env[62552]: DEBUG nova.compute.utils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.818548] env[62552]: DEBUG nova.compute.manager [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 817.818774] env[62552]: DEBUG nova.network.neutron [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.879630] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239377, 'name': CreateVM_Task, 'duration_secs': 0.680803} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.879935] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.880564] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.880746] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.881098] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.881422] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95dc0e7c-d9bf-4b8b-8738-d1c07ea840da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.886503] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 817.886503] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52112b9e-c965-effd-70e1-4a771481790f" [ 817.886503] env[62552]: _type = "Task" [ 817.886503] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.891782] env[62552]: DEBUG nova.policy [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12d1227fd454d3a9036c2e6c110a85f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd93b575f2e04fc9b93319545fa03708', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.898310] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52112b9e-c965-effd-70e1-4a771481790f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.937147] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af698514-df5a-47fa-a903-e16650c4665b tempest-ServerShowV254Test-504278264 tempest-ServerShowV254Test-504278264-project-member] Lock "1472ca01-ba04-4e7d-a9cf-ccaf29c04282" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.934s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.065996] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239373, 'name': CloneVM_Task} progress is 95%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.101666] env[62552]: DEBUG nova.network.neutron [req-401809a2-ad32-4970-8b09-9e925f82f9cf req-3951c57a-acfe-4a96-a185-69bfa8a173c1 service nova] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Updated VIF entry in instance network info cache for port 8cf42b44-1622-449e-8a17-e2be57435ee8. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 818.102077] env[62552]: DEBUG nova.network.neutron [req-401809a2-ad32-4970-8b09-9e925f82f9cf req-3951c57a-acfe-4a96-a185-69bfa8a173c1 service nova] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Updating instance_info_cache with network_info: [{"id": "8cf42b44-1622-449e-8a17-e2be57435ee8", "address": "fa:16:3e:c3:f5:7e", "network": {"id": "f5ce996d-7e56-4f9f-a7c2-f1adbc556126", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-601296890-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f9aa2ab2de4effbcbc7ea69c6db7dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cf42b44-16", "ovs_interfaceid": "8cf42b44-1622-449e-8a17-e2be57435ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.176264] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239381, 'name': ReconfigVM_Task, 'duration_secs': 0.273304} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.176264] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Reconfigured VM instance instance-0000002f to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 818.182629] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1462e9eb-18d4-4911-91a3-397744d52834 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.214981] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239382, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.218322] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 818.218322] env[62552]: value = "task-1239383" [ 818.218322] env[62552]: _type = "Task" [ 818.218322] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.233785] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239383, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.240306] env[62552]: DEBUG nova.network.neutron [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Successfully created port: 61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.268957] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239379, 'name': ReconfigVM_Task, 'duration_secs': 0.530461} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.269298] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 3b333977-0b4b-4ccf-bed5-231215abae3c/3b333977-0b4b-4ccf-bed5-231215abae3c.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.270760] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b3aedd9-f342-4a37-92ef-675b732d055e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.280079] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 818.280079] env[62552]: value = "task-1239384" [ 818.280079] env[62552]: _type = "Task" [ 818.280079] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.291819] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239384, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.320185] env[62552]: DEBUG nova.compute.manager [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 818.396470] env[62552]: DEBUG nova.network.neutron [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Successfully updated port: 76bce90f-8d24-4e1e-8562-f2790b183627 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 818.406510] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52112b9e-c965-effd-70e1-4a771481790f, 'name': SearchDatastore_Task, 'duration_secs': 0.035004} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.406510] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.406510] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.406510] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.406510] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.406510] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.406510] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-91025616-f53e-4213-b075-428a5504298a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.425531] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.425930] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 818.431163] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee70b36d-4932-4dd6-ae41-c648dfeb2ac4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.442114] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 818.442114] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b72565-e04f-153b-cd58-003a741534d0" [ 818.442114] env[62552]: _type = "Task" [ 818.442114] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.462357] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b72565-e04f-153b-cd58-003a741534d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.505691] env[62552]: DEBUG nova.compute.manager [req-9b6adc8f-f595-4469-a429-6b2f510333ca req-02e2db55-abad-4f30-af0a-c23bab5fa371 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Received event network-vif-plugged-76bce90f-8d24-4e1e-8562-f2790b183627 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 818.505966] env[62552]: DEBUG oslo_concurrency.lockutils [req-9b6adc8f-f595-4469-a429-6b2f510333ca req-02e2db55-abad-4f30-af0a-c23bab5fa371 service nova] Acquiring lock "80478878-ff82-4ed6-a851-8eb2bec01e22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.506319] env[62552]: DEBUG oslo_concurrency.lockutils [req-9b6adc8f-f595-4469-a429-6b2f510333ca req-02e2db55-abad-4f30-af0a-c23bab5fa371 service nova] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.506399] env[62552]: DEBUG oslo_concurrency.lockutils [req-9b6adc8f-f595-4469-a429-6b2f510333ca req-02e2db55-abad-4f30-af0a-c23bab5fa371 service nova] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.506583] env[62552]: DEBUG nova.compute.manager [req-9b6adc8f-f595-4469-a429-6b2f510333ca req-02e2db55-abad-4f30-af0a-c23bab5fa371 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] No waiting events found dispatching network-vif-plugged-76bce90f-8d24-4e1e-8562-f2790b183627 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 818.506752] env[62552]: WARNING nova.compute.manager [req-9b6adc8f-f595-4469-a429-6b2f510333ca req-02e2db55-abad-4f30-af0a-c23bab5fa371 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Received unexpected event network-vif-plugged-76bce90f-8d24-4e1e-8562-f2790b183627 for instance with vm_state building and task_state spawning. [ 818.568860] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239373, 'name': CloneVM_Task, 'duration_secs': 1.837185} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.569231] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Created linked-clone VM from snapshot [ 818.570105] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238cd112-e52e-44a7-acbe-292a710e1f37 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.580483] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Uploading image d88505cd-62b9-4da3-926e-217d15bab510 {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 818.604937] env[62552]: DEBUG oslo_concurrency.lockutils [req-401809a2-ad32-4970-8b09-9e925f82f9cf req-3951c57a-acfe-4a96-a185-69bfa8a173c1 service nova] Releasing lock "refresh_cache-fef94e30-f946-4d7f-a108-f32cad505bc4" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.611475] env[62552]: DEBUG oslo_vmware.rw_handles [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 818.611475] env[62552]: value = "vm-267398" [ 818.611475] env[62552]: _type = "VirtualMachine" [ 818.611475] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 818.611868] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6ebc2ace-ad9b-489b-ace5-4e9e424d27f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.623719] env[62552]: DEBUG oslo_vmware.rw_handles [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lease: (returnval){ [ 818.623719] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b66546-7a4d-89a1-8569-4b50bf0f5ec2" [ 818.623719] env[62552]: _type = "HttpNfcLease" [ 818.623719] env[62552]: } obtained for exporting VM: (result){ [ 818.623719] env[62552]: value = "vm-267398" [ 818.623719] env[62552]: _type = "VirtualMachine" [ 818.623719] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 818.624087] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the lease: (returnval){ [ 818.624087] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b66546-7a4d-89a1-8569-4b50bf0f5ec2" [ 818.624087] env[62552]: _type = "HttpNfcLease" [ 818.624087] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 818.635010] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 818.635010] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b66546-7a4d-89a1-8569-4b50bf0f5ec2" [ 818.635010] env[62552]: _type = "HttpNfcLease" [ 818.635010] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 818.716281] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239382, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.904611} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.716698] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 36233c83-1170-4a7a-be0b-6a0e8b139de7/36233c83-1170-4a7a-be0b-6a0e8b139de7.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 818.717564] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.717564] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3785083e-33c9-4cb8-a213-414cdce492a2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.733739] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239383, 'name': ReconfigVM_Task, 'duration_secs': 0.241099} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.735508] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267350', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'name': 'volume-0feb401f-606c-4b63-8c60-46d9d717225f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8707ce4b-677e-4f13-86f8-3e327d19380b', 'attached_at': '', 'detached_at': '', 'volume_id': '0feb401f-606c-4b63-8c60-46d9d717225f', 'serial': '0feb401f-606c-4b63-8c60-46d9d717225f'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 818.735966] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.736434] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 818.736434] env[62552]: value = "task-1239386" [ 818.736434] env[62552]: _type = "Task" [ 818.736434] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.739907] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed42196-16c4-43c3-9ff2-b146f5ff9a0a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.756975] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.764227] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 818.764993] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b205937-bb8d-4e8d-9f62-78e26ea31fb3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.791483] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239384, 'name': Rename_Task, 'duration_secs': 0.466316} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.791832] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 818.792039] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5c2615a-78cd-4e2b-8439-570bfb617f2b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.803270] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 818.803270] env[62552]: value = "task-1239388" [ 818.803270] env[62552]: _type = "Task" [ 818.803270] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.812759] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239388, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.837869] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33df3d67-3b65-4cc9-aaf6-c7a08503a191 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.849641] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08590b45-b6d7-476a-98b0-56e283f80f99 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.856254] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 818.856493] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 818.856714] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Deleting the datastore file [datastore2] 8707ce4b-677e-4f13-86f8-3e327d19380b {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 818.857633] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c684e65e-9025-4b3b-b37d-7492dcf20b12 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.892512] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b571d83d-7890-4778-8ea9-2b7b16b2bb6b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.896391] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for the task: (returnval){ [ 818.896391] env[62552]: value = "task-1239389" [ 818.896391] env[62552]: _type = "Task" [ 818.896391] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.906511] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "refresh_cache-80478878-ff82-4ed6-a851-8eb2bec01e22" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.906649] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquired lock "refresh_cache-80478878-ff82-4ed6-a851-8eb2bec01e22" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.906794] env[62552]: DEBUG nova.network.neutron [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 818.909701] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8eb891-2b8e-4a91-bf64-0fdaacbaf5e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.917353] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239389, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.930243] env[62552]: DEBUG nova.compute.provider_tree [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 818.954450] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b72565-e04f-153b-cd58-003a741534d0, 'name': SearchDatastore_Task, 'duration_secs': 0.06412} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.955341] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e335dd8c-37df-42b1-aae0-b040062c8041 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.962659] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 818.962659] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5213397e-16a1-aa07-84fd-fab166ce72d1" [ 818.962659] env[62552]: _type = "Task" [ 818.962659] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.971726] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5213397e-16a1-aa07-84fd-fab166ce72d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.133894] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 819.133894] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b66546-7a4d-89a1-8569-4b50bf0f5ec2" [ 819.133894] env[62552]: _type = "HttpNfcLease" [ 819.133894] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 819.134082] env[62552]: DEBUG oslo_vmware.rw_handles [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 819.134082] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b66546-7a4d-89a1-8569-4b50bf0f5ec2" [ 819.134082] env[62552]: _type = "HttpNfcLease" [ 819.134082] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 819.135117] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7872db-cdc2-4e06-ba2f-b0bc18558c77 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.143497] env[62552]: DEBUG oslo_vmware.rw_handles [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b824b1-d766-ec09-0451-25653caaef4b/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 819.143708] env[62552]: DEBUG oslo_vmware.rw_handles [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b824b1-d766-ec09-0451-25653caaef4b/disk-0.vmdk for reading. {{(pid=62552) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 819.255550] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.14247} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.255850] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.256666] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9ed6a9-19a6-479f-aed0-6c8b1d10e3b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.282207] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 36233c83-1170-4a7a-be0b-6a0e8b139de7/36233c83-1170-4a7a-be0b-6a0e8b139de7.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.284827] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40cd5a7d-7785-4c82-9d2f-05707b267e6c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.299844] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-35b8a206-1b56-42dd-ba86-e7d931689342 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.309109] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 819.309109] env[62552]: value = "task-1239390" [ 819.309109] env[62552]: _type = "Task" [ 819.309109] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.317875] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239388, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.326085] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239390, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.344828] env[62552]: DEBUG nova.compute.manager [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 819.409522] env[62552]: DEBUG oslo_vmware.api [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Task: {'id': task-1239389, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.260615} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.409522] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.409522] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 819.409747] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 819.452446] env[62552]: DEBUG nova.network.neutron [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.464595] env[62552]: ERROR nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [req-06a5c0f3-6a42-4be4-b65b-bda41140e25b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-06a5c0f3-6a42-4be4-b65b-bda41140e25b"}]} [ 819.490984] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5213397e-16a1-aa07-84fd-fab166ce72d1, 'name': SearchDatastore_Task, 'duration_secs': 0.028995} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.490984] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.491373] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] fef94e30-f946-4d7f-a108-f32cad505bc4/fef94e30-f946-4d7f-a108-f32cad505bc4.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 819.492048] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f2dfc586-3293-443d-aa99-aaecc44e8afb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.495295] env[62552]: DEBUG nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 819.505707] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 819.505707] env[62552]: value = "task-1239391" [ 819.505707] env[62552]: _type = "Task" [ 819.505707] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.513026] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 819.513489] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eedafbd5-86a7-49a5-bbe8-bb0e59081b97 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.523433] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239391, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.524618] env[62552]: DEBUG nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 819.524943] env[62552]: DEBUG nova.compute.provider_tree [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 819.538552] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b5450b-215f-4110-9f7e-d73933bf04ed {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.552699] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.553050] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.553273] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.553520] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.553726] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.553910] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.554199] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.554421] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.554906] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.554906] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.555025] env[62552]: DEBUG nova.virt.hardware [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.557752] env[62552]: DEBUG nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 819.559326] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f801436c-0562-42c5-b957-f7c94aed2dd5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.569203] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7df6e45-f4c4-4656-b43a-5a97d74e23b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.591159] env[62552]: ERROR nova.compute.manager [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Failed to detach volume 0feb401f-606c-4b63-8c60-46d9d717225f from /dev/sda: nova.exception.InstanceNotFound: Instance 8707ce4b-677e-4f13-86f8-3e327d19380b could not be found. [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Traceback (most recent call last): [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self.driver.rebuild(**kwargs) [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] raise NotImplementedError() [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] NotImplementedError [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] During handling of the above exception, another exception occurred: [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Traceback (most recent call last): [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self.driver.detach_volume(context, old_connection_info, [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] return self._volumeops.detach_volume(connection_info, instance) [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self._detach_volume_vmdk(connection_info, instance) [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] stable_ref.fetch_moref(session) [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] raise exception.InstanceNotFound(instance_id=self._uuid) [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] nova.exception.InstanceNotFound: Instance 8707ce4b-677e-4f13-86f8-3e327d19380b could not be found. [ 819.591159] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] [ 819.598242] env[62552]: DEBUG nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 819.730972] env[62552]: DEBUG nova.network.neutron [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Updating instance_info_cache with network_info: [{"id": "76bce90f-8d24-4e1e-8562-f2790b183627", "address": "fa:16:3e:c2:1e:92", "network": {"id": "1657d27d-1308-40b2-be9a-a1dd2b1da7ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1466254829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adb0b1256dd34e33913d5ae5bb3553b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76bce90f-8d", "ovs_interfaceid": "76bce90f-8d24-4e1e-8562-f2790b183627", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.813668] env[62552]: DEBUG nova.compute.manager [req-9179f1f2-49c5-49b3-aa15-9a72d5b61c52 req-dfa5e637-9fc1-466e-b6d8-483958e162f5 service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Received event network-vif-plugged-61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 819.814124] env[62552]: DEBUG oslo_concurrency.lockutils [req-9179f1f2-49c5-49b3-aa15-9a72d5b61c52 req-dfa5e637-9fc1-466e-b6d8-483958e162f5 service nova] Acquiring lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.814921] env[62552]: DEBUG oslo_concurrency.lockutils [req-9179f1f2-49c5-49b3-aa15-9a72d5b61c52 req-dfa5e637-9fc1-466e-b6d8-483958e162f5 service nova] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.814921] env[62552]: DEBUG oslo_concurrency.lockutils [req-9179f1f2-49c5-49b3-aa15-9a72d5b61c52 req-dfa5e637-9fc1-466e-b6d8-483958e162f5 service nova] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.814921] env[62552]: DEBUG nova.compute.manager [req-9179f1f2-49c5-49b3-aa15-9a72d5b61c52 req-dfa5e637-9fc1-466e-b6d8-483958e162f5 service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] No waiting events found dispatching network-vif-plugged-61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 819.815094] env[62552]: WARNING nova.compute.manager [req-9179f1f2-49c5-49b3-aa15-9a72d5b61c52 req-dfa5e637-9fc1-466e-b6d8-483958e162f5 service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Received unexpected event network-vif-plugged-61602401-8f5b-4b30-8d2d-b326bb568984 for instance with vm_state building and task_state spawning. [ 819.831386] env[62552]: DEBUG oslo_vmware.api [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239388, 'name': PowerOnVM_Task, 'duration_secs': 0.836577} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.831904] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 819.832045] env[62552]: INFO nova.compute.manager [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Took 8.11 seconds to spawn the instance on the hypervisor. [ 819.832149] env[62552]: DEBUG nova.compute.manager [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 819.834264] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91298950-5711-4369-8504-324903bb2623 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.842622] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239390, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.912136] env[62552]: DEBUG nova.compute.utils [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Build of instance 8707ce4b-677e-4f13-86f8-3e327d19380b aborted: Failed to rebuild volume backed instance. {{(pid=62552) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 819.918589] env[62552]: ERROR nova.compute.manager [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 8707ce4b-677e-4f13-86f8-3e327d19380b aborted: Failed to rebuild volume backed instance. [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Traceback (most recent call last): [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self.driver.rebuild(**kwargs) [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] raise NotImplementedError() [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] NotImplementedError [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] During handling of the above exception, another exception occurred: [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Traceback (most recent call last): [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 3642, in _rebuild_volume_backed_instance [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self._detach_root_volume(context, instance, root_bdm) [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 3621, in _detach_root_volume [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] with excutils.save_and_reraise_exception(): [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self.force_reraise() [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] raise self.value [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self.driver.detach_volume(context, old_connection_info, [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] return self._volumeops.detach_volume(connection_info, instance) [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self._detach_volume_vmdk(connection_info, instance) [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] stable_ref.fetch_moref(session) [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] raise exception.InstanceNotFound(instance_id=self._uuid) [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] nova.exception.InstanceNotFound: Instance 8707ce4b-677e-4f13-86f8-3e327d19380b could not be found. [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] During handling of the above exception, another exception occurred: [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Traceback (most recent call last): [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 11208, in _error_out_instance_on_exception [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] yield [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 3910, in rebuild_instance [ 819.918589] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self._do_rebuild_instance_with_claim( [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 3996, in _do_rebuild_instance_with_claim [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self._do_rebuild_instance( [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 4188, in _do_rebuild_instance [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self._rebuild_default_impl(**kwargs) [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 3765, in _rebuild_default_impl [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] self._rebuild_volume_backed_instance( [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] File "/opt/stack/nova/nova/compute/manager.py", line 3657, in _rebuild_volume_backed_instance [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] raise exception.BuildAbortException( [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] nova.exception.BuildAbortException: Build of instance 8707ce4b-677e-4f13-86f8-3e327d19380b aborted: Failed to rebuild volume backed instance. [ 819.919687] env[62552]: ERROR nova.compute.manager [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] [ 819.980134] env[62552]: DEBUG nova.network.neutron [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Successfully updated port: 61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.024837] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239391, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.125388] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727481dc-f4cb-44a1-8728-430601090289 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.144484] env[62552]: DEBUG oslo_vmware.rw_handles [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221cf72-750d-6c1e-7a7d-f40405ac2262/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 820.146727] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eae0b8d-2ec7-4bb7-ac8d-1f7e7b16769f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.152314] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a39a283-93f4-4b5d-b999-d95fec2b827a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.162594] env[62552]: DEBUG oslo_vmware.rw_handles [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221cf72-750d-6c1e-7a7d-f40405ac2262/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 820.164808] env[62552]: ERROR oslo_vmware.rw_handles [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221cf72-750d-6c1e-7a7d-f40405ac2262/disk-0.vmdk due to incomplete transfer. [ 820.199995] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e011eff2-7687-4073-bb00-3901119f6ade {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.203975] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da71d20-7f65-4a03-b23d-5b8b9eab388a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.217537] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0866a51-c87b-4c19-80e7-206f4c02c5ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.222406] env[62552]: DEBUG oslo_vmware.rw_handles [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221cf72-750d-6c1e-7a7d-f40405ac2262/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 820.222769] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Uploaded image 35ac675b-d35f-462e-aaa2-a2061f578780 to the Glance image server {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 820.225022] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 820.225367] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2675d552-eeed-4d76-ac50-7feb83adb31f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.239150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Releasing lock "refresh_cache-80478878-ff82-4ed6-a851-8eb2bec01e22" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.239689] env[62552]: DEBUG nova.compute.manager [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Instance network_info: |[{"id": "76bce90f-8d24-4e1e-8562-f2790b183627", "address": "fa:16:3e:c2:1e:92", "network": {"id": "1657d27d-1308-40b2-be9a-a1dd2b1da7ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1466254829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adb0b1256dd34e33913d5ae5bb3553b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76bce90f-8d", "ovs_interfaceid": "76bce90f-8d24-4e1e-8562-f2790b183627", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 820.240300] env[62552]: DEBUG nova.compute.provider_tree [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 820.245693] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:1e:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06cc7c49-c46c-4c1e-bf51-77e9ea802c40', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76bce90f-8d24-4e1e-8562-f2790b183627', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.251083] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Creating folder: Project (adb0b1256dd34e33913d5ae5bb3553b5). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.251583] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 820.251583] env[62552]: value = "task-1239392" [ 820.251583] env[62552]: _type = "Task" [ 820.251583] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.255367] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-999f5b07-939e-474c-bdf0-ab6b95b893db {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.266302] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239392, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.268230] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Created folder: Project (adb0b1256dd34e33913d5ae5bb3553b5) in parent group-v267339. [ 820.268634] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Creating folder: Instances. Parent ref: group-v267402. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.268943] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4fc01ea-578f-4635-a561-1a007c248d21 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.275318] env[62552]: ERROR nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [req-ac2db1c0-aee5-4ec0-9cf0-6b06b38e2443] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ac2db1c0-aee5-4ec0-9cf0-6b06b38e2443"}]} [ 820.280182] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Created folder: Instances in parent group-v267402. [ 820.280881] env[62552]: DEBUG oslo.service.loopingcall [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.280881] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 820.281377] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c96717a4-b402-4fba-ad51-4a80addba068 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.301401] env[62552]: DEBUG nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 820.309495] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.309495] env[62552]: value = "task-1239395" [ 820.309495] env[62552]: _type = "Task" [ 820.309495] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.320292] env[62552]: DEBUG nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 820.320292] env[62552]: DEBUG nova.compute.provider_tree [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 820.327142] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239395, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.327525] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239390, 'name': ReconfigVM_Task, 'duration_secs': 0.548542} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.327525] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 36233c83-1170-4a7a-be0b-6a0e8b139de7/36233c83-1170-4a7a-be0b-6a0e8b139de7.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.328593] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c34b55e-e6ea-4ca3-8cd9-fae0ae66ca52 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.336714] env[62552]: DEBUG nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 820.340679] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 820.340679] env[62552]: value = "task-1239396" [ 820.340679] env[62552]: _type = "Task" [ 820.340679] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.351546] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239396, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.368024] env[62552]: DEBUG nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 820.372354] env[62552]: INFO nova.compute.manager [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Took 34.33 seconds to build instance. [ 820.485607] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.485607] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquired lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.485607] env[62552]: DEBUG nova.network.neutron [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.520808] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239391, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572538} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.521143] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] fef94e30-f946-4d7f-a108-f32cad505bc4/fef94e30-f946-4d7f-a108-f32cad505bc4.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 820.521362] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.521644] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ff3f629-af7a-4f12-a4c0-92b0f0d6e407 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.530012] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 820.530012] env[62552]: value = "task-1239397" [ 820.530012] env[62552]: _type = "Task" [ 820.530012] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.542794] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239397, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.742872] env[62552]: DEBUG nova.compute.manager [req-4dd99e53-b6f7-4768-8fec-87cf0c48e250 req-127ea977-7404-4acd-985c-1330d05136a8 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Received event network-changed-76bce90f-8d24-4e1e-8562-f2790b183627 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 820.743094] env[62552]: DEBUG nova.compute.manager [req-4dd99e53-b6f7-4768-8fec-87cf0c48e250 req-127ea977-7404-4acd-985c-1330d05136a8 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Refreshing instance network info cache due to event network-changed-76bce90f-8d24-4e1e-8562-f2790b183627. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 820.743283] env[62552]: DEBUG oslo_concurrency.lockutils [req-4dd99e53-b6f7-4768-8fec-87cf0c48e250 req-127ea977-7404-4acd-985c-1330d05136a8 service nova] Acquiring lock "refresh_cache-80478878-ff82-4ed6-a851-8eb2bec01e22" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.743454] env[62552]: DEBUG oslo_concurrency.lockutils [req-4dd99e53-b6f7-4768-8fec-87cf0c48e250 req-127ea977-7404-4acd-985c-1330d05136a8 service nova] Acquired lock "refresh_cache-80478878-ff82-4ed6-a851-8eb2bec01e22" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.743697] env[62552]: DEBUG nova.network.neutron [req-4dd99e53-b6f7-4768-8fec-87cf0c48e250 req-127ea977-7404-4acd-985c-1330d05136a8 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Refreshing network info cache for port 76bce90f-8d24-4e1e-8562-f2790b183627 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.769889] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239392, 'name': Destroy_Task, 'duration_secs': 0.361005} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.770258] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Destroyed the VM [ 820.770687] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 820.771048] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-faea9561-0943-47ce-895e-20aa86d21480 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.781781] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 820.781781] env[62552]: value = "task-1239398" [ 820.781781] env[62552]: _type = "Task" [ 820.781781] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.787842] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b0e233-c60e-405d-9950-247da80ee536 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.802170] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239398, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.802170] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f232bb-5a31-4328-a9b6-2e28e5bc643b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.842559] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8a3e70-faee-4f53-949d-7089ee2b884d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.855367] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239395, 'name': CreateVM_Task, 'duration_secs': 0.47853} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.861104] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 820.861627] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239396, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.862482] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.862867] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.863482] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 820.865129] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bee856-6d61-443b-aafc-bdd4af1f2c89 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.869375] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6c52956-e873-474d-bcd1-a874c02f4c40 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.884171] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12022dd1-b382-41ce-a8ac-6589a82cae94 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.208s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.884171] env[62552]: DEBUG nova.compute.provider_tree [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 820.889049] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 820.889049] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52548173-ecc2-f421-229b-95e983df7da9" [ 820.889049] env[62552]: _type = "Task" [ 820.889049] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.900596] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52548173-ecc2-f421-229b-95e983df7da9, 'name': SearchDatastore_Task, 'duration_secs': 0.019742} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.901863] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.902261] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.902518] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.902703] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.903344] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.904868] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a298d69c-a99a-490a-9b18-c1259fc2847c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.913857] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.914488] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 820.915125] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f1cb1cd-9825-49b1-8b41-b2e4cdacc9e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.920703] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 820.920703] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f5ff8a-8dea-076f-0150-ec8059d46b9a" [ 820.920703] env[62552]: _type = "Task" [ 820.920703] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.929554] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f5ff8a-8dea-076f-0150-ec8059d46b9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.040993] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239397, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079587} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.042057] env[62552]: DEBUG nova.network.neutron [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.043995] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.045061] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4985ce01-88aa-401a-9c14-48b51c12e0bd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.070161] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] fef94e30-f946-4d7f-a108-f32cad505bc4/fef94e30-f946-4d7f-a108-f32cad505bc4.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.070672] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-918c8846-fe08-44f2-8d93-8cd9ab4819e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.094564] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 821.094564] env[62552]: value = "task-1239399" [ 821.094564] env[62552]: _type = "Task" [ 821.094564] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.103850] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239399, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.294962] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239398, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.328104] env[62552]: DEBUG nova.network.neutron [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Updating instance_info_cache with network_info: [{"id": "61602401-8f5b-4b30-8d2d-b326bb568984", "address": "fa:16:3e:06:29:86", "network": {"id": "3391f646-fd85-4a2a-b4a1-5d92620ed41b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1518992046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd93b575f2e04fc9b93319545fa03708", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61602401-8f", "ovs_interfaceid": "61602401-8f5b-4b30-8d2d-b326bb568984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.358901] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239396, 'name': Rename_Task, 'duration_secs': 0.584429} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.358901] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 821.360189] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-311e825d-e125-45e5-8210-2569beba242c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.372053] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 821.372053] env[62552]: value = "task-1239400" [ 821.372053] env[62552]: _type = "Task" [ 821.372053] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.384029] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239400, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.393789] env[62552]: DEBUG nova.compute.manager [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 821.433351] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f5ff8a-8dea-076f-0150-ec8059d46b9a, 'name': SearchDatastore_Task, 'duration_secs': 0.014022} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.433770] env[62552]: DEBUG nova.scheduler.client.report [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 78 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 821.436015] env[62552]: DEBUG nova.compute.provider_tree [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 78 to 79 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 821.436015] env[62552]: DEBUG nova.compute.provider_tree [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 821.444413] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f561a68-06d4-459d-9ee2-7279a29dbff7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.455035] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 821.455035] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5242679d-2ce2-ee12-d8f7-0c2f99d9be66" [ 821.455035] env[62552]: _type = "Task" [ 821.455035] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.467547] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5242679d-2ce2-ee12-d8f7-0c2f99d9be66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.611817] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.633765] env[62552]: DEBUG nova.network.neutron [req-4dd99e53-b6f7-4768-8fec-87cf0c48e250 req-127ea977-7404-4acd-985c-1330d05136a8 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Updated VIF entry in instance network info cache for port 76bce90f-8d24-4e1e-8562-f2790b183627. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 821.633829] env[62552]: DEBUG nova.network.neutron [req-4dd99e53-b6f7-4768-8fec-87cf0c48e250 req-127ea977-7404-4acd-985c-1330d05136a8 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Updating instance_info_cache with network_info: [{"id": "76bce90f-8d24-4e1e-8562-f2790b183627", "address": "fa:16:3e:c2:1e:92", "network": {"id": "1657d27d-1308-40b2-be9a-a1dd2b1da7ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1466254829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adb0b1256dd34e33913d5ae5bb3553b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76bce90f-8d", "ovs_interfaceid": "76bce90f-8d24-4e1e-8562-f2790b183627", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.798273] env[62552]: DEBUG oslo_vmware.api [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239398, 'name': RemoveSnapshot_Task, 'duration_secs': 0.657665} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.800068] env[62552]: DEBUG oslo_concurrency.lockutils [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "3b333977-0b4b-4ccf-bed5-231215abae3c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.800692] env[62552]: DEBUG oslo_concurrency.lockutils [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.803049] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 821.803408] env[62552]: INFO nova.compute.manager [None req-96f69483-e4b1-4318-9141-74b600f693ad tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Took 15.23 seconds to snapshot the instance on the hypervisor. [ 821.831984] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Releasing lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.832461] env[62552]: DEBUG nova.compute.manager [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Instance network_info: |[{"id": "61602401-8f5b-4b30-8d2d-b326bb568984", "address": "fa:16:3e:06:29:86", "network": {"id": "3391f646-fd85-4a2a-b4a1-5d92620ed41b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1518992046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd93b575f2e04fc9b93319545fa03708", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61602401-8f", "ovs_interfaceid": "61602401-8f5b-4b30-8d2d-b326bb568984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 821.833432] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:29:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '163e60bd-32d6-41c5-95e6-2eb10c5c9245', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61602401-8f5b-4b30-8d2d-b326bb568984', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.846947] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Creating folder: Project (bd93b575f2e04fc9b93319545fa03708). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.848156] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df9297c4-f793-4eca-a113-fe3e8fbb9a14 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.861830] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Created folder: Project (bd93b575f2e04fc9b93319545fa03708) in parent group-v267339. [ 821.861830] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Creating folder: Instances. Parent ref: group-v267405. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.861830] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06287179-bdc8-47a1-accb-47333a77cd69 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.871932] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Created folder: Instances in parent group-v267405. [ 821.872231] env[62552]: DEBUG oslo.service.loopingcall [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.872444] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.872761] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd798f97-4447-488b-b668-f02035524b03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.914820] env[62552]: DEBUG oslo_vmware.api [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239400, 'name': PowerOnVM_Task, 'duration_secs': 0.530543} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.915933] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 821.916372] env[62552]: INFO nova.compute.manager [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Took 12.69 seconds to spawn the instance on the hypervisor. [ 821.916617] env[62552]: DEBUG nova.compute.manager [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 821.917046] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.917046] env[62552]: value = "task-1239403" [ 821.917046] env[62552]: _type = "Task" [ 821.917046] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.918648] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad76a44a-28a2-415b-aade-61c6c081a637 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.941625] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.942631] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239403, 'name': CreateVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.944939] env[62552]: DEBUG oslo_concurrency.lockutils [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.945643] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.640s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.946240] env[62552]: DEBUG nova.compute.manager [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 821.953088] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.192s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.953088] env[62552]: DEBUG nova.objects.instance [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lazy-loading 'resources' on Instance uuid 00b06df7-e7ae-48fd-8887-65647931d5cf {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.969120] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5242679d-2ce2-ee12-d8f7-0c2f99d9be66, 'name': SearchDatastore_Task, 'duration_secs': 0.013434} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.969513] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.969846] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 80478878-ff82-4ed6-a851-8eb2bec01e22/80478878-ff82-4ed6-a851-8eb2bec01e22.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 821.970174] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a79ca56-d1ce-4442-abfb-be495c896543 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.977828] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 821.977828] env[62552]: value = "task-1239404" [ 821.977828] env[62552]: _type = "Task" [ 821.977828] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.991328] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239404, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.020284] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "526d0f25-3e99-4558-94c0-754ec2a80bad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.020359] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "526d0f25-3e99-4558-94c0-754ec2a80bad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.106170] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239399, 'name': ReconfigVM_Task, 'duration_secs': 0.607856} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.106497] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Reconfigured VM instance instance-00000039 to attach disk [datastore2] fef94e30-f946-4d7f-a108-f32cad505bc4/fef94e30-f946-4d7f-a108-f32cad505bc4.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.107161] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad311931-9380-4b1e-9e8e-924f2d17b11d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.114867] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 822.114867] env[62552]: value = "task-1239405" [ 822.114867] env[62552]: _type = "Task" [ 822.114867] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.121611] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239405, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.136973] env[62552]: DEBUG oslo_concurrency.lockutils [req-4dd99e53-b6f7-4768-8fec-87cf0c48e250 req-127ea977-7404-4acd-985c-1330d05136a8 service nova] Releasing lock "refresh_cache-80478878-ff82-4ed6-a851-8eb2bec01e22" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.307971] env[62552]: DEBUG nova.compute.utils [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.436799] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239403, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.453956] env[62552]: DEBUG nova.compute.utils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.456496] env[62552]: INFO nova.compute.manager [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Took 38.33 seconds to build instance. [ 822.457635] env[62552]: DEBUG nova.compute.manager [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 822.457884] env[62552]: DEBUG nova.network.neutron [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.490390] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239404, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.517460] env[62552]: DEBUG nova.compute.manager [req-5def0cc1-b3bd-483f-85d3-45184709225e req-f2b102c4-34da-43a6-9857-75453161152c service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Received event network-changed-61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 822.517610] env[62552]: DEBUG nova.compute.manager [req-5def0cc1-b3bd-483f-85d3-45184709225e req-f2b102c4-34da-43a6-9857-75453161152c service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Refreshing instance network info cache due to event network-changed-61602401-8f5b-4b30-8d2d-b326bb568984. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 822.517836] env[62552]: DEBUG oslo_concurrency.lockutils [req-5def0cc1-b3bd-483f-85d3-45184709225e req-f2b102c4-34da-43a6-9857-75453161152c service nova] Acquiring lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.517988] env[62552]: DEBUG oslo_concurrency.lockutils [req-5def0cc1-b3bd-483f-85d3-45184709225e req-f2b102c4-34da-43a6-9857-75453161152c service nova] Acquired lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.518198] env[62552]: DEBUG nova.network.neutron [req-5def0cc1-b3bd-483f-85d3-45184709225e req-f2b102c4-34da-43a6-9857-75453161152c service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Refreshing network info cache for port 61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.531144] env[62552]: DEBUG nova.policy [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bdea5d6f50634b5e89852744db56db67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8495bac65967441996ecca1fd22a4da4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.625822] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239405, 'name': Rename_Task, 'duration_secs': 0.17829} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.626302] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 822.628488] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6d26f39-53f1-4963-b709-34ace5418bea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.635123] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 822.635123] env[62552]: value = "task-1239406" [ 822.635123] env[62552]: _type = "Task" [ 822.635123] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.645072] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239406, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.812855] env[62552]: DEBUG oslo_concurrency.lockutils [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.907375] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae750d31-2cb6-46f7-95bc-d04225eefb38 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.917324] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7c051d-ae38-4a5f-8eb9-5e9d4087ad7d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.933657] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239403, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.961687] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb34a039-e684-454d-9961-e0a402e1b503 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.965150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-048bae9f-7a6d-4bf7-a0de-7a0aa84004fc tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.207s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.965662] env[62552]: DEBUG nova.compute.manager [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 822.976503] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984eaba9-4987-41a7-9ef8-78d4db4c35e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.983641] env[62552]: DEBUG nova.network.neutron [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Successfully created port: 43bb3381-583a-42f4-bcd9-3db694582a1d {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.998669] env[62552]: DEBUG nova.compute.provider_tree [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.007026] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239404, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.610492} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.007026] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 80478878-ff82-4ed6-a851-8eb2bec01e22/80478878-ff82-4ed6-a851-8eb2bec01e22.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 823.007026] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 823.007026] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f297a72c-fe51-49d1-945f-e5c225c5e053 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.011742] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 823.011742] env[62552]: value = "task-1239407" [ 823.011742] env[62552]: _type = "Task" [ 823.011742] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.023413] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239407, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.142800] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239406, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.297537] env[62552]: DEBUG nova.network.neutron [req-5def0cc1-b3bd-483f-85d3-45184709225e req-f2b102c4-34da-43a6-9857-75453161152c service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Updated VIF entry in instance network info cache for port 61602401-8f5b-4b30-8d2d-b326bb568984. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 823.298009] env[62552]: DEBUG nova.network.neutron [req-5def0cc1-b3bd-483f-85d3-45184709225e req-f2b102c4-34da-43a6-9857-75453161152c service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Updating instance_info_cache with network_info: [{"id": "61602401-8f5b-4b30-8d2d-b326bb568984", "address": "fa:16:3e:06:29:86", "network": {"id": "3391f646-fd85-4a2a-b4a1-5d92620ed41b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1518992046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd93b575f2e04fc9b93319545fa03708", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61602401-8f", "ovs_interfaceid": "61602401-8f5b-4b30-8d2d-b326bb568984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.442931] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239403, 'name': CreateVM_Task, 'duration_secs': 1.471839} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.443169] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 823.443904] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.444168] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.444508] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.444776] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f12bebd3-bbe6-4fa2-be98-3ae11329f5a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.450633] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 823.450633] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f5a08e-6db8-888d-dbdb-30e6a15fc853" [ 823.450633] env[62552]: _type = "Task" [ 823.450633] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.458882] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f5a08e-6db8-888d-dbdb-30e6a15fc853, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.483686] env[62552]: DEBUG nova.compute.manager [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 823.505519] env[62552]: DEBUG nova.scheduler.client.report [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 823.523157] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239407, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065043} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.524328] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.525201] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c79fc8-d35f-4f5d-bb90-b0b2161924d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.557219] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 80478878-ff82-4ed6-a851-8eb2bec01e22/80478878-ff82-4ed6-a851-8eb2bec01e22.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.557219] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-117c0cb6-5387-440f-bb44-6937b7b28f24 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.579034] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 823.579034] env[62552]: value = "task-1239408" [ 823.579034] env[62552]: _type = "Task" [ 823.579034] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.584567] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239408, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.646299] env[62552]: DEBUG oslo_vmware.api [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239406, 'name': PowerOnVM_Task, 'duration_secs': 0.875552} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.647662] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 823.649548] env[62552]: INFO nova.compute.manager [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Took 9.45 seconds to spawn the instance on the hypervisor. [ 823.649548] env[62552]: DEBUG nova.compute.manager [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 823.650382] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d5d086-1f7c-4d2c-9b78-f9511072f5e1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.806756] env[62552]: DEBUG oslo_concurrency.lockutils [req-5def0cc1-b3bd-483f-85d3-45184709225e req-f2b102c4-34da-43a6-9857-75453161152c service nova] Releasing lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.900880] env[62552]: DEBUG oslo_concurrency.lockutils [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "3b333977-0b4b-4ccf-bed5-231215abae3c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.901331] env[62552]: DEBUG oslo_concurrency.lockutils [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.901641] env[62552]: INFO nova.compute.manager [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Attaching volume 2f7588b6-1e37-4632-8f74-2950c19d197a to /dev/sdb [ 823.959171] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a578a9ed-bdb5-496a-af09-03356fd5f804 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.967224] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f5a08e-6db8-888d-dbdb-30e6a15fc853, 'name': SearchDatastore_Task, 'duration_secs': 0.020876} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.969755] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.970065] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.970335] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.970544] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.970729] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.975611] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa00e6ba-2228-4bf8-af9a-52d06b123d5f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.981320] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5380fa7f-58c0-4696-8d8f-3cace6c828cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.987274] env[62552]: DEBUG nova.compute.manager [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 823.996488] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.996724] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 823.997904] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cf100c1-2b5f-4f50-be5e-7be930878427 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.005215] env[62552]: DEBUG nova.virt.block_device [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Updating existing volume attachment record: 095cbaab-6fef-4649-94c6-58c2d7435d3d {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 824.013423] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.063s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.018520] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 824.018520] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dd7a9b-d03b-3484-3ed4-ee7737104baa" [ 824.018520] env[62552]: _type = "Task" [ 824.018520] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.018520] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.018520] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.018520] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.019133] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.019133] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.019133] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.019133] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.019259] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.019760] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.019760] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.020314] env[62552]: DEBUG nova.virt.hardware [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.020759] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.668s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.023117] env[62552]: INFO nova.compute.claims [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.029460] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc41fb88-3436-40b6-a69d-ba7834572c01 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.034505] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.046853] env[62552]: INFO nova.scheduler.client.report [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Deleted allocations for instance 00b06df7-e7ae-48fd-8887-65647931d5cf [ 824.051260] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2091fabe-2e88-4b4e-8d5e-62cdc4845ff1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.054273] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dd7a9b-d03b-3484-3ed4-ee7737104baa, 'name': SearchDatastore_Task, 'duration_secs': 0.021491} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.058765] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3408c302-73cd-4cad-8939-41be912cd114 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.073485] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 824.073485] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5292f6f5-8f9e-4ecb-51c2-d2c80737273b" [ 824.073485] env[62552]: _type = "Task" [ 824.073485] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.085365] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5292f6f5-8f9e-4ecb-51c2-d2c80737273b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.089796] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239408, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.135017] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "36233c83-1170-4a7a-be0b-6a0e8b139de7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.135017] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.135017] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "36233c83-1170-4a7a-be0b-6a0e8b139de7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.135218] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.135439] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.140926] env[62552]: INFO nova.compute.manager [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Terminating instance [ 824.175751] env[62552]: INFO nova.compute.manager [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Took 37.56 seconds to build instance. [ 824.482879] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Acquiring lock "8707ce4b-677e-4f13-86f8-3e327d19380b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.482981] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "8707ce4b-677e-4f13-86f8-3e327d19380b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.483405] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Acquiring lock "8707ce4b-677e-4f13-86f8-3e327d19380b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.483669] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "8707ce4b-677e-4f13-86f8-3e327d19380b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.483864] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "8707ce4b-677e-4f13-86f8-3e327d19380b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.486186] env[62552]: INFO nova.compute.manager [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Terminating instance [ 824.561351] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ee48aa9-af5b-408a-bca8-dc68437f2ff4 tempest-InstanceActionsTestJSON-74214199 tempest-InstanceActionsTestJSON-74214199-project-member] Lock "00b06df7-e7ae-48fd-8887-65647931d5cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.282s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.589335] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5292f6f5-8f9e-4ecb-51c2-d2c80737273b, 'name': SearchDatastore_Task, 'duration_secs': 0.014451} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.592704] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.593017] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28/df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 824.593614] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239408, 'name': ReconfigVM_Task, 'duration_secs': 0.596854} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.594161] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e388b3eb-7c8a-41ba-8351-c35af132caa1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.596633] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 80478878-ff82-4ed6-a851-8eb2bec01e22/80478878-ff82-4ed6-a851-8eb2bec01e22.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.597331] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-016cbb26-89a0-4635-a9ce-bf8c59bb7335 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.604022] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 824.604022] env[62552]: value = "task-1239412" [ 824.604022] env[62552]: _type = "Task" [ 824.604022] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.611332] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 824.611332] env[62552]: value = "task-1239413" [ 824.611332] env[62552]: _type = "Task" [ 824.611332] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.617999] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.625083] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239413, 'name': Rename_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.645700] env[62552]: DEBUG nova.compute.manager [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 824.645995] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 824.647035] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c708bd94-1dda-4b5a-8c40-ddf4553f766a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.655617] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 824.655933] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32d74891-8015-43f6-a16e-e9912c86e213 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.664296] env[62552]: DEBUG oslo_vmware.api [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 824.664296] env[62552]: value = "task-1239414" [ 824.664296] env[62552]: _type = "Task" [ 824.664296] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.671692] env[62552]: DEBUG oslo_vmware.api [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.680633] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70587b75-e048-49c7-834e-8031555f4a82 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "fef94e30-f946-4d7f-a108-f32cad505bc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.557s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.994646] env[62552]: DEBUG nova.compute.manager [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 824.995214] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0341d73e-e696-4202-876a-62759d8b9724 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.005990] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15401492-fbf2-42d2-b764-706e7c8cb90c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.044183] env[62552]: WARNING nova.virt.vmwareapi.driver [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 8707ce4b-677e-4f13-86f8-3e327d19380b could not be found. [ 825.044183] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 825.048027] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e6ccbc7-d8bf-4886-a933-173a747b575e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.054836] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0cd115-1142-4395-a2f6-90dbe1c848e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.086495] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8707ce4b-677e-4f13-86f8-3e327d19380b could not be found. [ 825.086755] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.087891] env[62552]: INFO nova.compute.manager [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Took 0.09 seconds to destroy the instance on the hypervisor. [ 825.087891] env[62552]: DEBUG oslo.service.loopingcall [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.091206] env[62552]: DEBUG nova.compute.manager [-] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 825.091206] env[62552]: DEBUG nova.network.neutron [-] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 825.115595] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239412, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50749} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.121574] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28/df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 825.121574] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.123033] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68d1f157-c908-4762-8001-5787b25a8022 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.133176] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239413, 'name': Rename_Task, 'duration_secs': 0.194807} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.134535] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 825.134864] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 825.134864] env[62552]: value = "task-1239415" [ 825.134864] env[62552]: _type = "Task" [ 825.134864] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.137686] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8fa3e263-78f7-4608-90cd-80bbbf691674 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.150145] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239415, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.151635] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 825.151635] env[62552]: value = "task-1239416" [ 825.151635] env[62552]: _type = "Task" [ 825.151635] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.162179] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.172598] env[62552]: DEBUG oslo_vmware.api [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239414, 'name': PowerOffVM_Task, 'duration_secs': 0.233442} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.175553] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 825.175869] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 825.176468] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ccdd5764-3a2d-42a7-b08b-f8d14f27114d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.183240] env[62552]: DEBUG nova.compute.manager [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 825.223841] env[62552]: DEBUG nova.network.neutron [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Successfully updated port: 43bb3381-583a-42f4-bcd9-3db694582a1d {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.280662] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 825.280662] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 825.280891] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Deleting the datastore file [datastore2] 36233c83-1170-4a7a-be0b-6a0e8b139de7 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.281490] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-818d936a-7e5a-47ee-9f3b-e107cf5a1d4e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.290027] env[62552]: DEBUG oslo_vmware.api [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for the task: (returnval){ [ 825.290027] env[62552]: value = "task-1239418" [ 825.290027] env[62552]: _type = "Task" [ 825.290027] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.301748] env[62552]: DEBUG oslo_vmware.api [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239418, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.452398] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceff6443-3b34-426d-81d7-3a711caca864 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.459856] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74711551-1c43-4faa-ad3b-e4afee770675 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.490474] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c69ac60-7434-48d7-9269-6da9a18573d7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.499126] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e50877-3a76-43d4-be76-ef5fb59fa960 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.515022] env[62552]: DEBUG nova.compute.provider_tree [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.522076] env[62552]: DEBUG nova.compute.manager [req-b97f9884-fc95-4adb-9558-fb8812549c19 req-845192f7-2925-4257-a1ba-10267ba6af7f service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Received event network-vif-plugged-43bb3381-583a-42f4-bcd9-3db694582a1d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 825.522463] env[62552]: DEBUG oslo_concurrency.lockutils [req-b97f9884-fc95-4adb-9558-fb8812549c19 req-845192f7-2925-4257-a1ba-10267ba6af7f service nova] Acquiring lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.522944] env[62552]: DEBUG oslo_concurrency.lockutils [req-b97f9884-fc95-4adb-9558-fb8812549c19 req-845192f7-2925-4257-a1ba-10267ba6af7f service nova] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.522944] env[62552]: DEBUG oslo_concurrency.lockutils [req-b97f9884-fc95-4adb-9558-fb8812549c19 req-845192f7-2925-4257-a1ba-10267ba6af7f service nova] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.523258] env[62552]: DEBUG nova.compute.manager [req-b97f9884-fc95-4adb-9558-fb8812549c19 req-845192f7-2925-4257-a1ba-10267ba6af7f service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] No waiting events found dispatching network-vif-plugged-43bb3381-583a-42f4-bcd9-3db694582a1d {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 825.523258] env[62552]: WARNING nova.compute.manager [req-b97f9884-fc95-4adb-9558-fb8812549c19 req-845192f7-2925-4257-a1ba-10267ba6af7f service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Received unexpected event network-vif-plugged-43bb3381-583a-42f4-bcd9-3db694582a1d for instance with vm_state building and task_state spawning. [ 825.653157] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239415, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067628} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.656586] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.658341] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4058454f-d5d1-4d8b-9ffd-9be77de0331a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.669026] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239416, 'name': PowerOnVM_Task} progress is 79%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.689901] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28/df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.692373] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3be08410-7f32-4eb1-9bae-58236cf8094e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.720953] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 825.720953] env[62552]: value = "task-1239419" [ 825.720953] env[62552]: _type = "Task" [ 825.720953] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.727691] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.732349] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.732628] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.732772] env[62552]: DEBUG nova.network.neutron [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.746252] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239419, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.802600] env[62552]: DEBUG oslo_vmware.api [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Task: {'id': task-1239418, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.450503} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.803914] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.804312] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 825.804548] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.804766] env[62552]: INFO nova.compute.manager [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 825.805039] env[62552]: DEBUG oslo.service.loopingcall [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.805350] env[62552]: DEBUG nova.compute.manager [-] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 825.805350] env[62552]: DEBUG nova.network.neutron [-] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 825.818771] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.819061] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.016025] env[62552]: DEBUG nova.scheduler.client.report [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 826.173334] env[62552]: DEBUG oslo_vmware.api [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239416, 'name': PowerOnVM_Task, 'duration_secs': 0.851787} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.173827] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 826.174137] env[62552]: INFO nova.compute.manager [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Took 9.41 seconds to spawn the instance on the hypervisor. [ 826.175909] env[62552]: DEBUG nova.compute.manager [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 826.175909] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cb7626-1e1e-4c85-bfce-46acd0b6d6bf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.235859] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Acquiring lock "ce5d0165-65f1-4505-9c46-1129c56a8913" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.236408] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "ce5d0165-65f1-4505-9c46-1129c56a8913" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.237155] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Acquiring lock "ce5d0165-65f1-4505-9c46-1129c56a8913-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.237155] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "ce5d0165-65f1-4505-9c46-1129c56a8913-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.237155] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "ce5d0165-65f1-4505-9c46-1129c56a8913-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.240958] env[62552]: INFO nova.compute.manager [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Terminating instance [ 826.250970] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239419, 'name': ReconfigVM_Task, 'duration_secs': 0.333424} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.250970] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Reconfigured VM instance instance-0000003b to attach disk [datastore1] df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28/df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.250970] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2cf51b98-238a-42ec-8b9e-7157ba48f14f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.258637] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 826.258637] env[62552]: value = "task-1239420" [ 826.258637] env[62552]: _type = "Task" [ 826.258637] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.274532] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239420, 'name': Rename_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.292713] env[62552]: DEBUG nova.network.neutron [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.312142] env[62552]: DEBUG nova.compute.manager [req-5384eef2-6cef-4258-996f-1349f8b612f2 req-bdc126b4-0fc9-4201-9e47-84e9344d0510 service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Received event network-vif-deleted-65c6abb5-29b3-4603-b019-4716db17392c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 826.312142] env[62552]: INFO nova.compute.manager [req-5384eef2-6cef-4258-996f-1349f8b612f2 req-bdc126b4-0fc9-4201-9e47-84e9344d0510 service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Neutron deleted interface 65c6abb5-29b3-4603-b019-4716db17392c; detaching it from the instance and deleting it from the info cache [ 826.312142] env[62552]: DEBUG nova.network.neutron [req-5384eef2-6cef-4258-996f-1349f8b612f2 req-bdc126b4-0fc9-4201-9e47-84e9344d0510 service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.522836] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.523435] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 826.537327] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.574s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.537327] env[62552]: INFO nova.compute.claims [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.628248] env[62552]: DEBUG nova.network.neutron [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.668023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquiring lock "fef94e30-f946-4d7f-a108-f32cad505bc4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.668023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "fef94e30-f946-4d7f-a108-f32cad505bc4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.668023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquiring lock "fef94e30-f946-4d7f-a108-f32cad505bc4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.668023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "fef94e30-f946-4d7f-a108-f32cad505bc4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.668023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "fef94e30-f946-4d7f-a108-f32cad505bc4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.669660] env[62552]: INFO nova.compute.manager [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Terminating instance [ 826.700158] env[62552]: INFO nova.compute.manager [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Took 38.48 seconds to build instance. [ 826.717391] env[62552]: DEBUG nova.network.neutron [-] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.746513] env[62552]: DEBUG nova.compute.manager [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 826.746754] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 826.747383] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f116b9b-2db2-4267-8878-94d37ae6aba6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.762023] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for the task: (returnval){ [ 826.762023] env[62552]: value = "task-1239422" [ 826.762023] env[62552]: _type = "Task" [ 826.762023] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.777055] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239420, 'name': Rename_Task, 'duration_secs': 0.185649} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.782192] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.782192] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239422, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.782192] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-457cb334-952e-4b66-9494-b2309b8a086c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.792180] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 826.792180] env[62552]: value = "task-1239423" [ 826.792180] env[62552]: _type = "Task" [ 826.792180] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.806326] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239423, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.818187] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a1c8493f-309b-4258-8843-66dbfda690e1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.830140] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6dc316c-b52b-46ca-8e9d-31b2ffc0af45 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.871692] env[62552]: DEBUG nova.compute.manager [req-5384eef2-6cef-4258-996f-1349f8b612f2 req-bdc126b4-0fc9-4201-9e47-84e9344d0510 service nova] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Detach interface failed, port_id=65c6abb5-29b3-4603-b019-4716db17392c, reason: Instance 8707ce4b-677e-4f13-86f8-3e327d19380b could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 827.043126] env[62552]: DEBUG nova.compute.utils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.049273] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 827.049273] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 827.133850] env[62552]: DEBUG nova.policy [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02fad99a3a1f453aada7427baa061ba3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abb4ff37a7154d7da41e936b8358f93d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.136016] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.136016] env[62552]: DEBUG nova.compute.manager [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Instance network_info: |[{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 827.136579] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:1c:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43bb3381-583a-42f4-bcd9-3db694582a1d', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.144470] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Creating folder: Project (8495bac65967441996ecca1fd22a4da4). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.145625] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2e4b344e-ef71-4aa6-b5ff-2ff10140c7ad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.157074] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Created folder: Project (8495bac65967441996ecca1fd22a4da4) in parent group-v267339. [ 827.157317] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Creating folder: Instances. Parent ref: group-v267410. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.158862] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18c976b2-a7fc-40b4-98e8-d4873c641bb6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.168613] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Created folder: Instances in parent group-v267410. [ 827.168923] env[62552]: DEBUG oslo.service.loopingcall [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.169156] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.169386] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-944acd0a-10ad-4b58-add8-8b2e0265e3fd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.186248] env[62552]: DEBUG nova.compute.manager [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 827.186466] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 827.187984] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9f50f4-8de8-4dfe-9023-7ec0dce29c68 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.197511] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 827.199237] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95f7b3b6-12bc-4574-a2b6-32a6e528c4e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.204020] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.204020] env[62552]: value = "task-1239426" [ 827.204020] env[62552]: _type = "Task" [ 827.204020] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.206180] env[62552]: DEBUG oslo_concurrency.lockutils [None req-db7973fd-bdbf-483d-819a-a20ea07c1679 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.029s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.206376] env[62552]: DEBUG oslo_vmware.api [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 827.206376] env[62552]: value = "task-1239427" [ 827.206376] env[62552]: _type = "Task" [ 827.206376] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.222778] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239426, 'name': CreateVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.222778] env[62552]: INFO nova.compute.manager [-] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Took 2.13 seconds to deallocate network for instance. [ 827.228034] env[62552]: DEBUG oslo_vmware.api [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239427, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.281336] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239422, 'name': PowerOffVM_Task, 'duration_secs': 0.260212} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.281867] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 827.282524] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 827.282524] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267357', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'name': 'volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce5d0165-65f1-4505-9c46-1129c56a8913', 'attached_at': '', 'detached_at': '', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'serial': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 827.283608] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee82c94-2a45-4c53-aef7-0b4c39130cf3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.307430] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1638bd67-8f24-4bd3-b7a5-2420cf99763e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.319033] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239423, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.320844] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672d63f1-d7c7-4b10-bbf8-9387829caa2e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.345111] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c3c514-f9ca-4555-a299-4dd9550a1706 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.366627] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] The volume has not been displaced from its original location: [datastore2] volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed/volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 827.369987] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Reconfiguring VM instance instance-00000030 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 827.370494] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-960e2345-76a5-44c9-b9f8-0bb40139f378 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.388859] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for the task: (returnval){ [ 827.388859] env[62552]: value = "task-1239428" [ 827.388859] env[62552]: _type = "Task" [ 827.388859] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.397095] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239428, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.549118] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 827.622921] env[62552]: DEBUG nova.network.neutron [-] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.711240] env[62552]: DEBUG nova.compute.manager [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 827.720691] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239426, 'name': CreateVM_Task, 'duration_secs': 0.450513} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.721170] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 827.722319] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.722489] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.722860] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 827.725622] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-445e15c6-df07-49b3-81f8-058eaa133653 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.731695] env[62552]: DEBUG oslo_vmware.api [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239427, 'name': PowerOffVM_Task, 'duration_secs': 0.242464} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.732570] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 827.732683] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 827.732968] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0423431d-b25f-454e-99b4-6ccb9e021c23 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.735964] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 827.735964] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5269e92d-121a-89f5-0861-532b2e27889d" [ 827.735964] env[62552]: _type = "Task" [ 827.735964] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.745482] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5269e92d-121a-89f5-0861-532b2e27889d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.792041] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Successfully created port: 8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.821375] env[62552]: DEBUG oslo_vmware.api [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239423, 'name': PowerOnVM_Task, 'duration_secs': 0.721951} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.822263] env[62552]: INFO nova.compute.manager [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Took 0.60 seconds to detach 1 volumes for instance. [ 827.827345] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.827893] env[62552]: INFO nova.compute.manager [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Took 8.48 seconds to spawn the instance on the hypervisor. [ 827.828356] env[62552]: DEBUG nova.compute.manager [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 827.829552] env[62552]: DEBUG nova.compute.manager [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Deleting volume: 0feb401f-606c-4b63-8c60-46d9d717225f {{(pid=62552) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 827.832659] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0237fa6-78d7-4564-871c-1dcc6e83668d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.863926] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 827.863926] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 827.864267] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Deleting the datastore file [datastore2] fef94e30-f946-4d7f-a108-f32cad505bc4 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.865239] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-126ae94c-1f57-4522-89bc-2c6f196d3ebc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.873575] env[62552]: DEBUG oslo_vmware.api [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for the task: (returnval){ [ 827.873575] env[62552]: value = "task-1239430" [ 827.873575] env[62552]: _type = "Task" [ 827.873575] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.886357] env[62552]: DEBUG oslo_vmware.api [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239430, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.902865] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239428, 'name': ReconfigVM_Task, 'duration_secs': 0.255331} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.904691] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Reconfigured VM instance instance-00000030 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 827.913075] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4eb52a58-7fff-41d7-9a9f-27e85cde50d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.934106] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for the task: (returnval){ [ 827.934106] env[62552]: value = "task-1239431" [ 827.934106] env[62552]: _type = "Task" [ 827.934106] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.947850] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239431, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.976935] env[62552]: DEBUG nova.compute.manager [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Received event network-changed-43bb3381-583a-42f4-bcd9-3db694582a1d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 827.977262] env[62552]: DEBUG nova.compute.manager [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Refreshing instance network info cache due to event network-changed-43bb3381-583a-42f4-bcd9-3db694582a1d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 827.977496] env[62552]: DEBUG oslo_concurrency.lockutils [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] Acquiring lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.977815] env[62552]: DEBUG oslo_concurrency.lockutils [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] Acquired lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.977815] env[62552]: DEBUG nova.network.neutron [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Refreshing network info cache for port 43bb3381-583a-42f4-bcd9-3db694582a1d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.064834] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1982d65a-950d-462a-b275-ff0a7263d62e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.075793] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074ac21c-c1f9-4078-ad4c-62b6eb56b3ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.111815] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d447ff-5629-4f56-93a9-0f7d4579e32e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.120926] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d579c37-12e9-40a8-92ef-cbcc5c7a557f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.135791] env[62552]: INFO nova.compute.manager [-] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Took 2.33 seconds to deallocate network for instance. [ 828.136740] env[62552]: DEBUG nova.compute.provider_tree [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.218838] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "0418260a-aa27-4955-ab15-b180ec04f0b3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.219153] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.243417] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.247666] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5269e92d-121a-89f5-0861-532b2e27889d, 'name': SearchDatastore_Task, 'duration_secs': 0.016615} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.247955] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.248208] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.248475] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.248646] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.248835] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.249105] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ac2b4a7-d22b-439c-8b44-1900ddee0339 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.258673] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.258872] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.260355] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4c7c615-b578-4932-8bd8-9a152d877e0b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.267816] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 828.267816] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5275b342-b93a-e12e-7c75-dcffd4590bac" [ 828.267816] env[62552]: _type = "Task" [ 828.267816] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.276996] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5275b342-b93a-e12e-7c75-dcffd4590bac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.359928] env[62552]: INFO nova.compute.manager [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Took 40.05 seconds to build instance. [ 828.388105] env[62552]: DEBUG oslo_vmware.api [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Task: {'id': task-1239430, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.307763} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.388926] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 828.389202] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 828.389732] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 828.389732] env[62552]: INFO nova.compute.manager [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Took 1.20 seconds to destroy the instance on the hypervisor. [ 828.389904] env[62552]: DEBUG oslo.service.loopingcall [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.390742] env[62552]: DEBUG nova.compute.manager [-] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 828.391397] env[62552]: DEBUG nova.network.neutron [-] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 828.436370] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.447017] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239431, 'name': ReconfigVM_Task, 'duration_secs': 0.200297} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.447017] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267357', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'name': 'volume-1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce5d0165-65f1-4505-9c46-1129c56a8913', 'attached_at': '', 'detached_at': '', 'volume_id': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed', 'serial': '1e3c30b0-8285-44ec-9795-8dca10fcf1ed'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 828.447289] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 828.448119] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885edd23-614f-4b9f-a2bd-7d480e7c706e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.455897] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 828.456369] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67ec21ff-286a-4d89-8d51-0b903c01c6de {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.524581] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 828.524846] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 828.525090] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Deleting the datastore file [datastore2] ce5d0165-65f1-4505-9c46-1129c56a8913 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 828.525385] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00b62b50-4c2b-4816-96ca-2c30f732c50f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.534567] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for the task: (returnval){ [ 828.534567] env[62552]: value = "task-1239434" [ 828.534567] env[62552]: _type = "Task" [ 828.534567] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.544561] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239434, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.577116] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 828.583537] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 828.583747] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267409', 'volume_id': '2f7588b6-1e37-4632-8f74-2950c19d197a', 'name': 'volume-2f7588b6-1e37-4632-8f74-2950c19d197a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3b333977-0b4b-4ccf-bed5-231215abae3c', 'attached_at': '', 'detached_at': '', 'volume_id': '2f7588b6-1e37-4632-8f74-2950c19d197a', 'serial': '2f7588b6-1e37-4632-8f74-2950c19d197a'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 828.584873] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5708517c-77ec-45f1-bbf7-4e5ef600a987 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.605574] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4baaf37e-5f38-44c3-ab74-7769cdc1b243 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.636653] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] volume-2f7588b6-1e37-4632-8f74-2950c19d197a/volume-2f7588b6-1e37-4632-8f74-2950c19d197a.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.639353] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.639599] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.639789] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.640065] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.640155] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.640321] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.640587] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.640793] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.640995] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.641189] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.641384] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.641737] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29f0fccd-5eb1-4edb-bafb-a81eae899734 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.656881] env[62552]: DEBUG nova.scheduler.client.report [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 828.661785] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42ec3d1-a163-4abb-b342-9a7f19c8a283 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.669698] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.678318] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff89e43-1f6c-4ca1-8a53-cb694b96061e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.683159] env[62552]: DEBUG oslo_vmware.api [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 828.683159] env[62552]: value = "task-1239435" [ 828.683159] env[62552]: _type = "Task" [ 828.683159] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.706330] env[62552]: DEBUG oslo_vmware.api [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239435, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.731364] env[62552]: DEBUG nova.compute.utils [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.779677] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5275b342-b93a-e12e-7c75-dcffd4590bac, 'name': SearchDatastore_Task, 'duration_secs': 0.020045} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.782035] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-851ee7d3-6887-475f-a22b-4b5ff0e8dbe6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.786339] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 828.786339] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526ae267-284f-cd4a-eef8-3608750a4547" [ 828.786339] env[62552]: _type = "Task" [ 828.786339] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.794666] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526ae267-284f-cd4a-eef8-3608750a4547, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.862389] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e48b85ed-65c7-4e0f-a154-dda60694e884 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.878s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.976172] env[62552]: DEBUG nova.compute.manager [req-8beaaf69-e104-4f9a-a8ad-a38d47380168 req-b0356d4e-e8fd-4547-99ae-9015c09670c3 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Received event network-changed-76bce90f-8d24-4e1e-8562-f2790b183627 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 828.976172] env[62552]: DEBUG nova.compute.manager [req-8beaaf69-e104-4f9a-a8ad-a38d47380168 req-b0356d4e-e8fd-4547-99ae-9015c09670c3 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Refreshing instance network info cache due to event network-changed-76bce90f-8d24-4e1e-8562-f2790b183627. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 828.976172] env[62552]: DEBUG oslo_concurrency.lockutils [req-8beaaf69-e104-4f9a-a8ad-a38d47380168 req-b0356d4e-e8fd-4547-99ae-9015c09670c3 service nova] Acquiring lock "refresh_cache-80478878-ff82-4ed6-a851-8eb2bec01e22" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.976494] env[62552]: DEBUG oslo_concurrency.lockutils [req-8beaaf69-e104-4f9a-a8ad-a38d47380168 req-b0356d4e-e8fd-4547-99ae-9015c09670c3 service nova] Acquired lock "refresh_cache-80478878-ff82-4ed6-a851-8eb2bec01e22" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.976708] env[62552]: DEBUG nova.network.neutron [req-8beaaf69-e104-4f9a-a8ad-a38d47380168 req-b0356d4e-e8fd-4547-99ae-9015c09670c3 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Refreshing network info cache for port 76bce90f-8d24-4e1e-8562-f2790b183627 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 829.052266] env[62552]: DEBUG oslo_vmware.api [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Task: {'id': task-1239434, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135377} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.052266] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.052266] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 829.052266] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.052266] env[62552]: INFO nova.compute.manager [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Took 2.31 seconds to destroy the instance on the hypervisor. [ 829.053192] env[62552]: DEBUG oslo.service.loopingcall [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.054423] env[62552]: DEBUG nova.network.neutron [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updated VIF entry in instance network info cache for port 43bb3381-583a-42f4-bcd9-3db694582a1d. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 829.054648] env[62552]: DEBUG nova.network.neutron [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.056030] env[62552]: DEBUG nova.compute.manager [-] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 829.056030] env[62552]: DEBUG nova.network.neutron [-] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 829.169408] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.638s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.169971] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 829.173335] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.517s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.174807] env[62552]: INFO nova.compute.claims [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.205187] env[62552]: DEBUG oslo_vmware.api [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239435, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.234560] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.015s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.302023] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526ae267-284f-cd4a-eef8-3608750a4547, 'name': SearchDatastore_Task, 'duration_secs': 0.022414} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.302369] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.302637] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 88d0f502-fc3f-429e-bdf6-a1ebacec2117/88d0f502-fc3f-429e-bdf6-a1ebacec2117.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 829.302929] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dff1d8a6-06c1-4e55-b19d-9c620e58594a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.311853] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 829.311853] env[62552]: value = "task-1239436" [ 829.311853] env[62552]: _type = "Task" [ 829.311853] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.327552] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239436, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.365592] env[62552]: DEBUG nova.compute.manager [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 829.511212] env[62552]: DEBUG nova.network.neutron [-] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.552264] env[62552]: DEBUG oslo_vmware.rw_handles [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b824b1-d766-ec09-0451-25653caaef4b/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 829.556914] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85325516-ce0d-485e-8427-37890d681031 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.560967] env[62552]: DEBUG oslo_concurrency.lockutils [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] Releasing lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.560967] env[62552]: DEBUG nova.compute.manager [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Received event network-vif-deleted-fc44bf04-e20d-419e-a8ca-04c24a9fcf2e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 829.561276] env[62552]: DEBUG nova.compute.manager [req-b16f9a88-b955-40df-89eb-f564b5db81be req-cb2b9612-0b18-4bf1-9563-85ef79c2ae22 service nova] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Received event network-vif-deleted-9de2eca6-381f-431c-9e6f-b422ec21ac62 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 829.561784] env[62552]: DEBUG oslo_vmware.rw_handles [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b824b1-d766-ec09-0451-25653caaef4b/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 829.562048] env[62552]: ERROR oslo_vmware.rw_handles [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b824b1-d766-ec09-0451-25653caaef4b/disk-0.vmdk due to incomplete transfer. [ 829.562367] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-628b5acb-b76b-487f-b71f-93fe90bfce07 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.574032] env[62552]: DEBUG oslo_vmware.rw_handles [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b824b1-d766-ec09-0451-25653caaef4b/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 829.574032] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Uploaded image d88505cd-62b9-4da3-926e-217d15bab510 to the Glance image server {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 829.574032] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 829.574563] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-74fc32f6-7ffe-4496-b394-b87d7f4259b6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.581966] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 829.581966] env[62552]: value = "task-1239437" [ 829.581966] env[62552]: _type = "Task" [ 829.581966] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.590309] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239437, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.679935] env[62552]: DEBUG nova.compute.utils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.684719] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 829.685391] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.700832] env[62552]: DEBUG oslo_vmware.api [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239435, 'name': ReconfigVM_Task, 'duration_secs': 0.65619} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.701131] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Reconfigured VM instance instance-00000038 to attach disk [datastore1] volume-2f7588b6-1e37-4632-8f74-2950c19d197a/volume-2f7588b6-1e37-4632-8f74-2950c19d197a.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.706689] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9ae0cb8-90ae-47ac-ac15-d3f74b7b18d7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.724188] env[62552]: DEBUG oslo_vmware.api [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 829.724188] env[62552]: value = "task-1239438" [ 829.724188] env[62552]: _type = "Task" [ 829.724188] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.733604] env[62552]: DEBUG oslo_vmware.api [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239438, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.827874] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239436, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.876917] env[62552]: DEBUG nova.policy [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02fad99a3a1f453aada7427baa061ba3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abb4ff37a7154d7da41e936b8358f93d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.901740] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.015407] env[62552]: INFO nova.compute.manager [-] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Took 1.62 seconds to deallocate network for instance. [ 830.095561] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239437, 'name': Destroy_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.184187] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 830.237178] env[62552]: DEBUG oslo_vmware.api [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239438, 'name': ReconfigVM_Task, 'duration_secs': 0.20929} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.240814] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267409', 'volume_id': '2f7588b6-1e37-4632-8f74-2950c19d197a', 'name': 'volume-2f7588b6-1e37-4632-8f74-2950c19d197a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3b333977-0b4b-4ccf-bed5-231215abae3c', 'attached_at': '', 'detached_at': '', 'volume_id': '2f7588b6-1e37-4632-8f74-2950c19d197a', 'serial': '2f7588b6-1e37-4632-8f74-2950c19d197a'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 830.316264] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "0418260a-aa27-4955-ab15-b180ec04f0b3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.316546] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.317345] env[62552]: INFO nova.compute.manager [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Attaching volume 2d394306-c3fc-458b-bd15-0d1b8649225e to /dev/sdb [ 830.335280] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239436, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.88003} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.335381] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 88d0f502-fc3f-429e-bdf6-a1ebacec2117/88d0f502-fc3f-429e-bdf6-a1ebacec2117.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.335591] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.335859] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f7342d5-ada9-4fef-831a-692619551572 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.345579] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 830.345579] env[62552]: value = "task-1239439" [ 830.345579] env[62552]: _type = "Task" [ 830.345579] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.365247] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239439, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.376601] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcd8f5e-c06a-4fa1-aa34-fe09b0a8c2a1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.386313] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa13588-d2a3-46f1-9218-32bef88ed2dd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.401215] env[62552]: DEBUG nova.virt.block_device [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updating existing volume attachment record: caa2b9cd-3cc0-4973-a66b-583b6011956f {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 830.427708] env[62552]: DEBUG nova.network.neutron [req-8beaaf69-e104-4f9a-a8ad-a38d47380168 req-b0356d4e-e8fd-4547-99ae-9015c09670c3 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Updated VIF entry in instance network info cache for port 76bce90f-8d24-4e1e-8562-f2790b183627. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 830.428114] env[62552]: DEBUG nova.network.neutron [req-8beaaf69-e104-4f9a-a8ad-a38d47380168 req-b0356d4e-e8fd-4547-99ae-9015c09670c3 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Updating instance_info_cache with network_info: [{"id": "76bce90f-8d24-4e1e-8562-f2790b183627", "address": "fa:16:3e:c2:1e:92", "network": {"id": "1657d27d-1308-40b2-be9a-a1dd2b1da7ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1466254829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adb0b1256dd34e33913d5ae5bb3553b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76bce90f-8d", "ovs_interfaceid": "76bce90f-8d24-4e1e-8562-f2790b183627", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.527408] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.582937] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Successfully updated port: 8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 830.598469] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239437, 'name': Destroy_Task, 'duration_secs': 0.697539} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.599056] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Destroyed the VM [ 830.599167] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 830.599333] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b9a11e17-c45b-4fcd-a0db-63bab30b4175 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.608069] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 830.608069] env[62552]: value = "task-1239441" [ 830.608069] env[62552]: _type = "Task" [ 830.608069] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.616604] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239441, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.636755] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8482717-3a94-4979-9817-a0cb07bc3235 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.648938] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1027e3-429f-4471-8593-6136208a3d00 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.683870] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f362171a-e05e-425e-8252-aa08ad2740df {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.698104] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8decb124-8dee-40cd-a09b-7a9040a306b6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.706257] env[62552]: DEBUG nova.network.neutron [-] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.716044] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Successfully created port: 810c9b00-6705-4e67-b4f6-f711f07f2037 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.726161] env[62552]: DEBUG nova.compute.provider_tree [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.799730] env[62552]: DEBUG nova.compute.manager [req-88f0822f-cbc4-4c31-9ecd-040e33ede618 req-4029b7f9-d5c3-42d8-858d-ee19324705d4 service nova] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Received event network-vif-plugged-8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 830.799967] env[62552]: DEBUG oslo_concurrency.lockutils [req-88f0822f-cbc4-4c31-9ecd-040e33ede618 req-4029b7f9-d5c3-42d8-858d-ee19324705d4 service nova] Acquiring lock "474dfa9e-d003-478f-b48f-09e0c0452ffd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.800194] env[62552]: DEBUG oslo_concurrency.lockutils [req-88f0822f-cbc4-4c31-9ecd-040e33ede618 req-4029b7f9-d5c3-42d8-858d-ee19324705d4 service nova] Lock "474dfa9e-d003-478f-b48f-09e0c0452ffd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.800365] env[62552]: DEBUG oslo_concurrency.lockutils [req-88f0822f-cbc4-4c31-9ecd-040e33ede618 req-4029b7f9-d5c3-42d8-858d-ee19324705d4 service nova] Lock "474dfa9e-d003-478f-b48f-09e0c0452ffd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.800782] env[62552]: DEBUG nova.compute.manager [req-88f0822f-cbc4-4c31-9ecd-040e33ede618 req-4029b7f9-d5c3-42d8-858d-ee19324705d4 service nova] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] No waiting events found dispatching network-vif-plugged-8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 830.800841] env[62552]: WARNING nova.compute.manager [req-88f0822f-cbc4-4c31-9ecd-040e33ede618 req-4029b7f9-d5c3-42d8-858d-ee19324705d4 service nova] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Received unexpected event network-vif-plugged-8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f for instance with vm_state building and task_state spawning. [ 830.857032] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239439, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066401} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.857411] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.858323] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ce6038-c47d-4e99-a8a4-ef80f0d7f65e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.888271] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 88d0f502-fc3f-429e-bdf6-a1ebacec2117/88d0f502-fc3f-429e-bdf6-a1ebacec2117.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.888697] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b8220d9-9911-444d-a8c7-8b10a178a727 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.911799] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 830.911799] env[62552]: value = "task-1239444" [ 830.911799] env[62552]: _type = "Task" [ 830.911799] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.919729] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239444, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.932656] env[62552]: DEBUG oslo_concurrency.lockutils [req-8beaaf69-e104-4f9a-a8ad-a38d47380168 req-b0356d4e-e8fd-4547-99ae-9015c09670c3 service nova] Releasing lock "refresh_cache-80478878-ff82-4ed6-a851-8eb2bec01e22" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.016670] env[62552]: DEBUG nova.compute.manager [req-e4167841-35ad-49cf-8db1-efe747c9ca06 req-dcb95405-07dd-4fc0-92d3-bdc1343db103 service nova] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Received event network-vif-deleted-8cf42b44-1622-449e-8a17-e2be57435ee8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 831.086751] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "refresh_cache-474dfa9e-d003-478f-b48f-09e0c0452ffd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.086956] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "refresh_cache-474dfa9e-d003-478f-b48f-09e0c0452ffd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.087075] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.122578] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239441, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.208386] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 831.213115] env[62552]: INFO nova.compute.manager [-] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Took 2.16 seconds to deallocate network for instance. [ 831.232683] env[62552]: DEBUG nova.scheduler.client.report [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 831.259017] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.259017] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.259337] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.259337] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.259472] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.259622] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.259833] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.259995] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.260883] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.260883] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.260883] env[62552]: DEBUG nova.virt.hardware [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.263062] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5aa4f5-38cf-4ae3-ba3d-60e3faa4be5d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.275937] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d736e9b-4df9-4e82-ba59-2f5dfa2b4e2d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.311539] env[62552]: DEBUG nova.objects.instance [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'flavor' on Instance uuid 3b333977-0b4b-4ccf-bed5-231215abae3c {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.331161] env[62552]: DEBUG nova.compute.manager [req-2a8fdace-fd35-4580-ada9-f4250058eca3 req-14e054c9-aa8f-4690-b877-e4769c217f3a service nova] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Received event network-changed-8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 831.331161] env[62552]: DEBUG nova.compute.manager [req-2a8fdace-fd35-4580-ada9-f4250058eca3 req-14e054c9-aa8f-4690-b877-e4769c217f3a service nova] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Refreshing instance network info cache due to event network-changed-8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 831.331161] env[62552]: DEBUG oslo_concurrency.lockutils [req-2a8fdace-fd35-4580-ada9-f4250058eca3 req-14e054c9-aa8f-4690-b877-e4769c217f3a service nova] Acquiring lock "refresh_cache-474dfa9e-d003-478f-b48f-09e0c0452ffd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.426990] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.619729] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239441, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.633608] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.734029] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.734586] env[62552]: DEBUG nova.compute.manager [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 831.741199] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.787s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.741313] env[62552]: DEBUG nova.objects.instance [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lazy-loading 'resources' on Instance uuid 4b80bc41-1b63-444e-8039-696e3d78d9a1 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.782113] env[62552]: INFO nova.compute.manager [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Took 0.57 seconds to detach 1 volumes for instance. [ 831.784607] env[62552]: DEBUG nova.compute.manager [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Deleting volume: 1e3c30b0-8285-44ec-9795-8dca10fcf1ed {{(pid=62552) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 831.819983] env[62552]: DEBUG oslo_concurrency.lockutils [None req-39f44ce4-edfa-4ce5-a184-8ae02acc38a3 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.919s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.925098] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239444, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.952253] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Updating instance_info_cache with network_info: [{"id": "8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f", "address": "fa:16:3e:8e:43:da", "network": {"id": "3341135c-a39c-4c4c-bc65-d60626d959cd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-815900522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abb4ff37a7154d7da41e936b8358f93d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "821e0800-fabf-48d0-87b4-db5a1eddce93", "external-id": "nsx-vlan-transportzone-97", "segmentation_id": 97, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e7f9aff-c6", "ovs_interfaceid": "8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.955776] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "3b333977-0b4b-4ccf-bed5-231215abae3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.955776] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.955776] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "3b333977-0b4b-4ccf-bed5-231215abae3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.956824] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.956824] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.959940] env[62552]: INFO nova.compute.manager [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Terminating instance [ 832.123267] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239441, 'name': RemoveSnapshot_Task, 'duration_secs': 1.494979} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.123267] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 832.123267] env[62552]: DEBUG nova.compute.manager [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 832.123910] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7122f8-dedc-433b-b6dc-e0b00a71fd3b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.128788] env[62552]: DEBUG oslo_concurrency.lockutils [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.128788] env[62552]: DEBUG oslo_concurrency.lockutils [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.128788] env[62552]: INFO nova.compute.manager [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Rebooting instance [ 832.244385] env[62552]: DEBUG nova.compute.utils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.251857] env[62552]: DEBUG nova.compute.manager [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 832.252063] env[62552]: DEBUG nova.network.neutron [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 832.330486] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.345764] env[62552]: DEBUG nova.policy [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c732a4a15f9b4b16b241c09f4ba373e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3daa5253a0cc4538903b7cabc2f98fa2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.429707] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239444, 'name': ReconfigVM_Task, 'duration_secs': 1.055878} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.429707] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 88d0f502-fc3f-429e-bdf6-a1ebacec2117/88d0f502-fc3f-429e-bdf6-a1ebacec2117.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.429919] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-490fbff8-d6cb-491d-a537-9bee5aa13e00 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.438625] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 832.438625] env[62552]: value = "task-1239446" [ 832.438625] env[62552]: _type = "Task" [ 832.438625] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.456924] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "refresh_cache-474dfa9e-d003-478f-b48f-09e0c0452ffd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.456924] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Instance network_info: |[{"id": "8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f", "address": "fa:16:3e:8e:43:da", "network": {"id": "3341135c-a39c-4c4c-bc65-d60626d959cd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-815900522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abb4ff37a7154d7da41e936b8358f93d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "821e0800-fabf-48d0-87b4-db5a1eddce93", "external-id": "nsx-vlan-transportzone-97", "segmentation_id": 97, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e7f9aff-c6", "ovs_interfaceid": "8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 832.456924] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239446, 'name': Rename_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.456924] env[62552]: DEBUG oslo_concurrency.lockutils [req-2a8fdace-fd35-4580-ada9-f4250058eca3 req-14e054c9-aa8f-4690-b877-e4769c217f3a service nova] Acquired lock "refresh_cache-474dfa9e-d003-478f-b48f-09e0c0452ffd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.456924] env[62552]: DEBUG nova.network.neutron [req-2a8fdace-fd35-4580-ada9-f4250058eca3 req-14e054c9-aa8f-4690-b877-e4769c217f3a service nova] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Refreshing network info cache for port 8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.458990] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:43:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '821e0800-fabf-48d0-87b4-db5a1eddce93', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.473020] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Creating folder: Project (abb4ff37a7154d7da41e936b8358f93d). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 832.477856] env[62552]: DEBUG nova.compute.manager [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 832.478100] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.478657] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e77ff14-867e-4d1f-9ab5-fff040914fcd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.481830] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-721ba9af-7dd1-41bd-9ea2-c122d5646c60 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.489211] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 832.489211] env[62552]: value = "task-1239448" [ 832.489211] env[62552]: _type = "Task" [ 832.489211] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.498438] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Created folder: Project (abb4ff37a7154d7da41e936b8358f93d) in parent group-v267339. [ 832.498717] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Creating folder: Instances. Parent ref: group-v267415. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 832.500155] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-848caa54-116c-4bda-8aa9-0419d89feade {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.506686] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239448, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.516029] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Created folder: Instances in parent group-v267415. [ 832.516029] env[62552]: DEBUG oslo.service.loopingcall [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.516029] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 832.516029] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ef16c5b-24fc-42a3-8d60-99fa4986b9ac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.540778] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.540778] env[62552]: value = "task-1239450" [ 832.540778] env[62552]: _type = "Task" [ 832.540778] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.550578] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239450, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.645226] env[62552]: INFO nova.compute.manager [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Shelve offloading [ 832.650669] env[62552]: DEBUG oslo_concurrency.lockutils [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.650669] env[62552]: DEBUG oslo_concurrency.lockutils [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquired lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.650853] env[62552]: DEBUG nova.network.neutron [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.737467] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78a6210-c4b9-4407-9688-330d38c8a569 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.745725] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b9a7dc-029b-42f9-bc9a-c35354a78421 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.784961] env[62552]: DEBUG nova.compute.manager [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 832.787314] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd4383a-def9-452f-ab73-ff48d5fb121a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.796223] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e711bd04-348a-436b-8c46-f60d5b13b15c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.814143] env[62552]: DEBUG nova.compute.provider_tree [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.882019] env[62552]: DEBUG nova.network.neutron [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Successfully created port: f749d931-7a86-41ca-a551-4f8d2013603a {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.887485] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Successfully updated port: 810c9b00-6705-4e67-b4f6-f711f07f2037 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.951123] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239446, 'name': Rename_Task, 'duration_secs': 0.205586} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.951123] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.951123] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-21221e22-6f64-49c1-b63d-2b7a12f93f29 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.957040] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 832.957040] env[62552]: value = "task-1239452" [ 832.957040] env[62552]: _type = "Task" [ 832.957040] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.964415] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239452, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.003197] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239448, 'name': PowerOffVM_Task, 'duration_secs': 0.229132} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.003536] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.003812] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 833.004016] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267409', 'volume_id': '2f7588b6-1e37-4632-8f74-2950c19d197a', 'name': 'volume-2f7588b6-1e37-4632-8f74-2950c19d197a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3b333977-0b4b-4ccf-bed5-231215abae3c', 'attached_at': '', 'detached_at': '', 'volume_id': '2f7588b6-1e37-4632-8f74-2950c19d197a', 'serial': '2f7588b6-1e37-4632-8f74-2950c19d197a'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 833.004818] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d12200-01d2-4d01-9c0b-49a73d35ea0d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.025500] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6633e2-671d-423d-a31d-9586659361ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.032679] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8457e983-3b84-4788-87f9-11cd3847f5c6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.056600] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b9d36f-c0d4-4c7b-baec-d1ac1405aeb5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.068892] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239450, 'name': CreateVM_Task, 'duration_secs': 0.404061} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.081171] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.081359] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] The volume has not been displaced from its original location: [datastore1] volume-2f7588b6-1e37-4632-8f74-2950c19d197a/volume-2f7588b6-1e37-4632-8f74-2950c19d197a.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 833.086557] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Reconfiguring VM instance instance-00000038 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 833.087347] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.089064] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.089064] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.089064] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78c2a51d-71d6-4854-9a7a-e193f8ae565d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.101134] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c055ae3-59a1-4747-a949-3c06bcbf3671 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.106700] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 833.106700] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522f5083-9a5d-a076-c633-f4e38509fa9f" [ 833.106700] env[62552]: _type = "Task" [ 833.106700] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.111736] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 833.111736] env[62552]: value = "task-1239453" [ 833.111736] env[62552]: _type = "Task" [ 833.111736] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.117713] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522f5083-9a5d-a076-c633-f4e38509fa9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.123128] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239453, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.157304] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.157304] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-369ecd98-1e96-4b2d-ad85-1567b46dd631 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.167902] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "11831580-1b58-476a-91ce-a4e55947fd91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.168184] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "11831580-1b58-476a-91ce-a4e55947fd91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.168496] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 833.168496] env[62552]: value = "task-1239454" [ 833.168496] env[62552]: _type = "Task" [ 833.168496] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.178540] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 833.178800] env[62552]: DEBUG nova.compute.manager [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 833.179587] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31fffc2-c5e7-4ada-9594-ed1db4ba5968 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.189900] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.189900] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.189900] env[62552]: DEBUG nova.network.neutron [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.317992] env[62552]: DEBUG nova.scheduler.client.report [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 833.392103] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "refresh_cache-c52efdeb-9efc-4ba1-b9af-4dd0a925517f" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.392103] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "refresh_cache-c52efdeb-9efc-4ba1-b9af-4dd0a925517f" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.396165] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.468061] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239452, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.605285] env[62552]: DEBUG nova.network.neutron [req-2a8fdace-fd35-4580-ada9-f4250058eca3 req-14e054c9-aa8f-4690-b877-e4769c217f3a service nova] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Updated VIF entry in instance network info cache for port 8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 833.605652] env[62552]: DEBUG nova.network.neutron [req-2a8fdace-fd35-4580-ada9-f4250058eca3 req-14e054c9-aa8f-4690-b877-e4769c217f3a service nova] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Updating instance_info_cache with network_info: [{"id": "8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f", "address": "fa:16:3e:8e:43:da", "network": {"id": "3341135c-a39c-4c4c-bc65-d60626d959cd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-815900522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abb4ff37a7154d7da41e936b8358f93d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "821e0800-fabf-48d0-87b4-db5a1eddce93", "external-id": "nsx-vlan-transportzone-97", "segmentation_id": 97, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e7f9aff-c6", "ovs_interfaceid": "8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.618906] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522f5083-9a5d-a076-c633-f4e38509fa9f, 'name': SearchDatastore_Task, 'duration_secs': 0.01092} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.620138] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.620548] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.620712] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.620897] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.621050] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.621715] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d470d39b-95e6-493a-b880-05ccc73cd78c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.626945] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239453, 'name': ReconfigVM_Task, 'duration_secs': 0.279242} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.627540] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Reconfigured VM instance instance-00000038 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 833.633024] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12851aef-217e-4279-9883-667b0028babb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.645713] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.645911] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 833.647187] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30dae767-84ef-46ff-b570-6bcb740aab4b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.652053] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 833.652053] env[62552]: value = "task-1239455" [ 833.652053] env[62552]: _type = "Task" [ 833.652053] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.653287] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 833.653287] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528f7358-5e3e-f17b-78c9-eb05f782fe20" [ 833.653287] env[62552]: _type = "Task" [ 833.653287] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.670528] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239455, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.674439] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528f7358-5e3e-f17b-78c9-eb05f782fe20, 'name': SearchDatastore_Task, 'duration_secs': 0.011193} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.675427] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f804499-9f56-430f-a6b9-eedcc1568b3b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.680938] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 833.680938] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52712f72-8d94-6a82-1ea9-e18d5118e3ba" [ 833.680938] env[62552]: _type = "Task" [ 833.680938] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.689246] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52712f72-8d94-6a82-1ea9-e18d5118e3ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.728018] env[62552]: DEBUG nova.network.neutron [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Updating instance_info_cache with network_info: [{"id": "61602401-8f5b-4b30-8d2d-b326bb568984", "address": "fa:16:3e:06:29:86", "network": {"id": "3391f646-fd85-4a2a-b4a1-5d92620ed41b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1518992046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd93b575f2e04fc9b93319545fa03708", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61602401-8f", "ovs_interfaceid": "61602401-8f5b-4b30-8d2d-b326bb568984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.797132] env[62552]: DEBUG nova.compute.manager [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Received event network-vif-deleted-5cb2de18-f6f3-4351-a156-891bb4a6c542 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 833.797132] env[62552]: DEBUG nova.compute.manager [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Received event network-changed-61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 833.797132] env[62552]: DEBUG nova.compute.manager [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Refreshing instance network info cache due to event network-changed-61602401-8f5b-4b30-8d2d-b326bb568984. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 833.797132] env[62552]: DEBUG oslo_concurrency.lockutils [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] Acquiring lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.797503] env[62552]: DEBUG nova.compute.manager [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 833.823581] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.086s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.828777] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.618s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.831233] env[62552]: INFO nova.compute.claims [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.842885] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.844482] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.844482] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.844482] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.844482] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.844482] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.844482] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.844482] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.844810] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.844810] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.844924] env[62552]: DEBUG nova.virt.hardware [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.845944] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4735e98c-d395-41f2-abdb-6980a0f95d8e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.858708] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7184ab-8c02-4789-998b-78f55391d3a1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.864458] env[62552]: INFO nova.scheduler.client.report [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Deleted allocations for instance 4b80bc41-1b63-444e-8039-696e3d78d9a1 [ 833.974128] env[62552]: DEBUG oslo_vmware.api [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239452, 'name': PowerOnVM_Task, 'duration_secs': 0.557646} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.974128] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.974474] env[62552]: INFO nova.compute.manager [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Took 9.99 seconds to spawn the instance on the hypervisor. [ 833.974746] env[62552]: DEBUG nova.compute.manager [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 833.975434] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e36f6533-2ab1-4795-b622-6ad389ab77e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.023117] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.113910] env[62552]: DEBUG oslo_concurrency.lockutils [req-2a8fdace-fd35-4580-ada9-f4250058eca3 req-14e054c9-aa8f-4690-b877-e4769c217f3a service nova] Releasing lock "refresh_cache-474dfa9e-d003-478f-b48f-09e0c0452ffd" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.166473] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239455, 'name': ReconfigVM_Task, 'duration_secs': 0.22109} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.167165] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267409', 'volume_id': '2f7588b6-1e37-4632-8f74-2950c19d197a', 'name': 'volume-2f7588b6-1e37-4632-8f74-2950c19d197a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3b333977-0b4b-4ccf-bed5-231215abae3c', 'attached_at': '', 'detached_at': '', 'volume_id': '2f7588b6-1e37-4632-8f74-2950c19d197a', 'serial': '2f7588b6-1e37-4632-8f74-2950c19d197a'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 834.167856] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.168870] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab377f3b-fbee-4a1d-b56d-cc35ff53532c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.181662] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 834.186854] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34ced525-eee5-49aa-a243-67b2cfaca6d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.195710] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52712f72-8d94-6a82-1ea9-e18d5118e3ba, 'name': SearchDatastore_Task, 'duration_secs': 0.0153} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.195710] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.195710] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 474dfa9e-d003-478f-b48f-09e0c0452ffd/474dfa9e-d003-478f-b48f-09e0c0452ffd.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 834.195710] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71e3cbf9-6ba8-42a0-85ce-8b3e9dad8565 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.203397] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 834.203397] env[62552]: value = "task-1239457" [ 834.203397] env[62552]: _type = "Task" [ 834.203397] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.213657] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239457, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.230787] env[62552]: DEBUG oslo_concurrency.lockutils [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Releasing lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.230787] env[62552]: DEBUG oslo_concurrency.lockutils [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] Acquired lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.230787] env[62552]: DEBUG nova.network.neutron [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Refreshing network info cache for port 61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.259164] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 834.259895] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 834.259895] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleting the datastore file [datastore2] 3b333977-0b4b-4ccf-bed5-231215abae3c {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.260098] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6621c29-cca0-44af-a8a2-2ac4297f04ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.267770] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 834.267770] env[62552]: value = "task-1239458" [ 834.267770] env[62552]: _type = "Task" [ 834.267770] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.279729] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239458, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.398397] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ef21359-f12b-4e77-bb35-85248340f0a9 tempest-ServerGroupTestJSON-2032345807 tempest-ServerGroupTestJSON-2032345807-project-member] Lock "4b80bc41-1b63-444e-8039-696e3d78d9a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.086s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.436449] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "6c8b5367-4fa0-479b-9382-ff261201e3ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.436449] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "6c8b5367-4fa0-479b-9382-ff261201e3ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.500116] env[62552]: INFO nova.compute.manager [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Took 44.69 seconds to build instance. [ 834.573865] env[62552]: DEBUG nova.network.neutron [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating instance_info_cache with network_info: [{"id": "929d55b0-5a35-4d2d-a172-15d79a215977", "address": "fa:16:3e:62:4d:9b", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap929d55b0-5a", "ovs_interfaceid": "929d55b0-5a35-4d2d-a172-15d79a215977", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.651936] env[62552]: DEBUG nova.network.neutron [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Updating instance_info_cache with network_info: [{"id": "810c9b00-6705-4e67-b4f6-f711f07f2037", "address": "fa:16:3e:a7:41:91", "network": {"id": "3341135c-a39c-4c4c-bc65-d60626d959cd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-815900522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abb4ff37a7154d7da41e936b8358f93d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "821e0800-fabf-48d0-87b4-db5a1eddce93", "external-id": "nsx-vlan-transportzone-97", "segmentation_id": 97, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap810c9b00-67", "ovs_interfaceid": "810c9b00-6705-4e67-b4f6-f711f07f2037", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.715291] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239457, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.735409] env[62552]: DEBUG nova.compute.manager [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 834.738470] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230b3081-7f31-469d-b9df-f2b7b7e6ee2a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.779861] env[62552]: DEBUG nova.network.neutron [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Successfully updated port: f749d931-7a86-41ca-a551-4f8d2013603a {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.784861] env[62552]: DEBUG oslo_vmware.api [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239458, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191409} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.786144] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.786144] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 834.786234] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.786435] env[62552]: INFO nova.compute.manager [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Took 2.31 seconds to destroy the instance on the hypervisor. [ 834.786741] env[62552]: DEBUG oslo.service.loopingcall [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.789331] env[62552]: DEBUG nova.compute.manager [-] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 834.789331] env[62552]: DEBUG nova.network.neutron [-] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.962892] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 834.963012] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267414', 'volume_id': '2d394306-c3fc-458b-bd15-0d1b8649225e', 'name': 'volume-2d394306-c3fc-458b-bd15-0d1b8649225e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0418260a-aa27-4955-ab15-b180ec04f0b3', 'attached_at': '', 'detached_at': '', 'volume_id': '2d394306-c3fc-458b-bd15-0d1b8649225e', 'serial': '2d394306-c3fc-458b-bd15-0d1b8649225e'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 834.964391] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bcfab6-7d56-4c49-9d81-50c653d11ab0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.990225] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b84b885-d9af-457f-8377-61b04a57dafe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.009206] env[62552]: DEBUG oslo_concurrency.lockutils [None req-83a73c5f-9b64-4514-8063-125ad4d644c8 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.626s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.016912] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] volume-2d394306-c3fc-458b-bd15-0d1b8649225e/volume-2d394306-c3fc-458b-bd15-0d1b8649225e.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.023242] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e088f5b8-2466-4d98-adf6-ab8cac2ae5c3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.043462] env[62552]: DEBUG oslo_vmware.api [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 835.043462] env[62552]: value = "task-1239459" [ 835.043462] env[62552]: _type = "Task" [ 835.043462] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.056764] env[62552]: DEBUG oslo_vmware.api [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239459, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.079645] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.156167] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "refresh_cache-c52efdeb-9efc-4ba1-b9af-4dd0a925517f" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.156167] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Instance network_info: |[{"id": "810c9b00-6705-4e67-b4f6-f711f07f2037", "address": "fa:16:3e:a7:41:91", "network": {"id": "3341135c-a39c-4c4c-bc65-d60626d959cd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-815900522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abb4ff37a7154d7da41e936b8358f93d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "821e0800-fabf-48d0-87b4-db5a1eddce93", "external-id": "nsx-vlan-transportzone-97", "segmentation_id": 97, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap810c9b00-67", "ovs_interfaceid": "810c9b00-6705-4e67-b4f6-f711f07f2037", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 835.156511] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:41:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '821e0800-fabf-48d0-87b4-db5a1eddce93', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '810c9b00-6705-4e67-b4f6-f711f07f2037', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.164012] env[62552]: DEBUG oslo.service.loopingcall [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.166194] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.166350] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e134abed-493a-4c11-9ca1-7b6ba4540246 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.194237] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.194237] env[62552]: value = "task-1239460" [ 835.194237] env[62552]: _type = "Task" [ 835.194237] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.204028] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239460, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.214825] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239457, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.593397} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.215117] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 474dfa9e-d003-478f-b48f-09e0c0452ffd/474dfa9e-d003-478f-b48f-09e0c0452ffd.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.215336] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.215598] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5055eb4e-9524-445c-8444-6759312fbabe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.222792] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 835.222792] env[62552]: value = "task-1239461" [ 835.222792] env[62552]: _type = "Task" [ 835.222792] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.235060] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239461, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.254969] env[62552]: DEBUG nova.network.neutron [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Updated VIF entry in instance network info cache for port 61602401-8f5b-4b30-8d2d-b326bb568984. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.255458] env[62552]: DEBUG nova.network.neutron [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Updating instance_info_cache with network_info: [{"id": "61602401-8f5b-4b30-8d2d-b326bb568984", "address": "fa:16:3e:06:29:86", "network": {"id": "3391f646-fd85-4a2a-b4a1-5d92620ed41b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1518992046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd93b575f2e04fc9b93319545fa03708", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61602401-8f", "ovs_interfaceid": "61602401-8f5b-4b30-8d2d-b326bb568984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.261991] env[62552]: DEBUG nova.compute.manager [req-85894568-16fa-4372-bc2b-d97921c28517 req-26069b5d-a9c4-4747-8a1c-99eff8f75ba2 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Received event network-vif-plugged-f749d931-7a86-41ca-a551-4f8d2013603a {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 835.262216] env[62552]: DEBUG oslo_concurrency.lockutils [req-85894568-16fa-4372-bc2b-d97921c28517 req-26069b5d-a9c4-4747-8a1c-99eff8f75ba2 service nova] Acquiring lock "22357d4e-9771-477c-9fc3-fe3d76f6e902-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.262429] env[62552]: DEBUG oslo_concurrency.lockutils [req-85894568-16fa-4372-bc2b-d97921c28517 req-26069b5d-a9c4-4747-8a1c-99eff8f75ba2 service nova] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.262602] env[62552]: DEBUG oslo_concurrency.lockutils [req-85894568-16fa-4372-bc2b-d97921c28517 req-26069b5d-a9c4-4747-8a1c-99eff8f75ba2 service nova] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.262775] env[62552]: DEBUG nova.compute.manager [req-85894568-16fa-4372-bc2b-d97921c28517 req-26069b5d-a9c4-4747-8a1c-99eff8f75ba2 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] No waiting events found dispatching network-vif-plugged-f749d931-7a86-41ca-a551-4f8d2013603a {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 835.263012] env[62552]: WARNING nova.compute.manager [req-85894568-16fa-4372-bc2b-d97921c28517 req-26069b5d-a9c4-4747-8a1c-99eff8f75ba2 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Received unexpected event network-vif-plugged-f749d931-7a86-41ca-a551-4f8d2013603a for instance with vm_state building and task_state spawning. [ 835.284601] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.284601] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.285031] env[62552]: DEBUG nova.network.neutron [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.376728] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00435206-a5df-4eac-acc0-872a32046c94 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.387277] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411505d7-d59e-4031-a8d2-22e9b4085307 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.419687] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04619e1-3031-4733-a4a5-6880e3fab075 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.427704] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44873bb9-2e7b-4ff9-a502-15d2a440262f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.442049] env[62552]: DEBUG nova.compute.provider_tree [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.537808] env[62552]: DEBUG nova.compute.manager [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 835.558020] env[62552]: DEBUG oslo_vmware.api [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239459, 'name': ReconfigVM_Task, 'duration_secs': 0.367626} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.558020] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Reconfigured VM instance instance-00000031 to attach disk [datastore2] volume-2d394306-c3fc-458b-bd15-0d1b8649225e/volume-2d394306-c3fc-458b-bd15-0d1b8649225e.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.566176] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e94dbc4b-98e5-475a-ac26-b6dd343e2838 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.588438] env[62552]: DEBUG oslo_vmware.api [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 835.588438] env[62552]: value = "task-1239462" [ 835.588438] env[62552]: _type = "Task" [ 835.588438] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.601446] env[62552]: DEBUG oslo_vmware.api [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239462, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.602930] env[62552]: DEBUG nova.network.neutron [-] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.707487] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239460, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.733546] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239461, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076287} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.733862] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.734650] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b944ea-1cc8-4604-9e1a-c69579331aa8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.757455] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 474dfa9e-d003-478f-b48f-09e0c0452ffd/474dfa9e-d003-478f-b48f-09e0c0452ffd.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.758340] env[62552]: DEBUG oslo_concurrency.lockutils [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] Releasing lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.758581] env[62552]: DEBUG nova.compute.manager [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Received event network-vif-plugged-810c9b00-6705-4e67-b4f6-f711f07f2037 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 835.758775] env[62552]: DEBUG oslo_concurrency.lockutils [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] Acquiring lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.758981] env[62552]: DEBUG oslo_concurrency.lockutils [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] Lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.759196] env[62552]: DEBUG oslo_concurrency.lockutils [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] Lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.759449] env[62552]: DEBUG nova.compute.manager [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] No waiting events found dispatching network-vif-plugged-810c9b00-6705-4e67-b4f6-f711f07f2037 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 835.759656] env[62552]: WARNING nova.compute.manager [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Received unexpected event network-vif-plugged-810c9b00-6705-4e67-b4f6-f711f07f2037 for instance with vm_state building and task_state spawning. [ 835.759834] env[62552]: DEBUG nova.compute.manager [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Received event network-changed-810c9b00-6705-4e67-b4f6-f711f07f2037 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 835.760994] env[62552]: DEBUG nova.compute.manager [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Refreshing instance network info cache due to event network-changed-810c9b00-6705-4e67-b4f6-f711f07f2037. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 835.760994] env[62552]: DEBUG oslo_concurrency.lockutils [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] Acquiring lock "refresh_cache-c52efdeb-9efc-4ba1-b9af-4dd0a925517f" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.760994] env[62552]: DEBUG oslo_concurrency.lockutils [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] Acquired lock "refresh_cache-c52efdeb-9efc-4ba1-b9af-4dd0a925517f" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.760994] env[62552]: DEBUG nova.network.neutron [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Refreshing network info cache for port 810c9b00-6705-4e67-b4f6-f711f07f2037 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.762266] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d981666-ca19-4b3f-bc3d-fe493f4b92e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.780243] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af249fd5-0626-48ea-9323-3b20048c82ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.788353] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Doing hard reboot of VM {{(pid=62552) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 835.791165] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-c5b750c2-6451-4cc1-b6c0-916cea825837 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.792739] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 835.792739] env[62552]: value = "task-1239463" [ 835.792739] env[62552]: _type = "Task" [ 835.792739] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.797533] env[62552]: DEBUG oslo_vmware.api [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 835.797533] env[62552]: value = "task-1239464" [ 835.797533] env[62552]: _type = "Task" [ 835.797533] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.806331] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239463, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.812052] env[62552]: DEBUG oslo_vmware.api [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239464, 'name': ResetVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.817399] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.818585] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffad61b-d286-4a1c-aa01-d4509a22e311 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.826773] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.827162] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-341b05f1-8d2a-47b5-865a-ab8ae1864967 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.829654] env[62552]: DEBUG nova.network.neutron [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.907759] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.907759] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.907759] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleting the datastore file [datastore2] cdee22a9-4327-47af-ab14-8403a06b2802 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.910336] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7e544d4-52fb-4efd-83b0-aaabc0258ba7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.916431] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 835.916431] env[62552]: value = "task-1239466" [ 835.916431] env[62552]: _type = "Task" [ 835.916431] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.925538] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239466, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.973018] env[62552]: ERROR nova.scheduler.client.report [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [req-bc56f289-a231-4188-b2e7-2b171d586a24] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bc56f289-a231-4188-b2e7-2b171d586a24"}]} [ 835.989349] env[62552]: DEBUG nova.scheduler.client.report [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 835.992663] env[62552]: DEBUG nova.network.neutron [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance_info_cache with network_info: [{"id": "f749d931-7a86-41ca-a551-4f8d2013603a", "address": "fa:16:3e:00:02:b8", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf749d931-7a", "ovs_interfaceid": "f749d931-7a86-41ca-a551-4f8d2013603a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.011734] env[62552]: DEBUG nova.scheduler.client.report [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 836.012148] env[62552]: DEBUG nova.compute.provider_tree [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 836.027260] env[62552]: DEBUG nova.scheduler.client.report [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 836.059778] env[62552]: DEBUG nova.scheduler.client.report [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 836.066112] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.068482] env[62552]: DEBUG nova.compute.manager [req-2bd0dd9a-583a-4f4e-a958-e4dcbb06186f req-0b94be95-19ed-4de1-b0b4-15777e46d9fa service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received event network-vif-unplugged-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 836.068737] env[62552]: DEBUG oslo_concurrency.lockutils [req-2bd0dd9a-583a-4f4e-a958-e4dcbb06186f req-0b94be95-19ed-4de1-b0b4-15777e46d9fa service nova] Acquiring lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.068965] env[62552]: DEBUG oslo_concurrency.lockutils [req-2bd0dd9a-583a-4f4e-a958-e4dcbb06186f req-0b94be95-19ed-4de1-b0b4-15777e46d9fa service nova] Lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.069175] env[62552]: DEBUG oslo_concurrency.lockutils [req-2bd0dd9a-583a-4f4e-a958-e4dcbb06186f req-0b94be95-19ed-4de1-b0b4-15777e46d9fa service nova] Lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.069350] env[62552]: DEBUG nova.compute.manager [req-2bd0dd9a-583a-4f4e-a958-e4dcbb06186f req-0b94be95-19ed-4de1-b0b4-15777e46d9fa service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] No waiting events found dispatching network-vif-unplugged-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 836.069529] env[62552]: WARNING nova.compute.manager [req-2bd0dd9a-583a-4f4e-a958-e4dcbb06186f req-0b94be95-19ed-4de1-b0b4-15777e46d9fa service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received unexpected event network-vif-unplugged-929d55b0-5a35-4d2d-a172-15d79a215977 for instance with vm_state shelved and task_state shelving_offloading. [ 836.108469] env[62552]: INFO nova.compute.manager [-] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Took 1.32 seconds to deallocate network for instance. [ 836.108898] env[62552]: DEBUG oslo_vmware.api [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239462, 'name': ReconfigVM_Task, 'duration_secs': 0.181488} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.112017] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267414', 'volume_id': '2d394306-c3fc-458b-bd15-0d1b8649225e', 'name': 'volume-2d394306-c3fc-458b-bd15-0d1b8649225e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0418260a-aa27-4955-ab15-b180ec04f0b3', 'attached_at': '', 'detached_at': '', 'volume_id': '2d394306-c3fc-458b-bd15-0d1b8649225e', 'serial': '2d394306-c3fc-458b-bd15-0d1b8649225e'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 836.158127] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquiring lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.158127] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.179841] env[62552]: DEBUG nova.network.neutron [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Updated VIF entry in instance network info cache for port 810c9b00-6705-4e67-b4f6-f711f07f2037. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.180505] env[62552]: DEBUG nova.network.neutron [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Updating instance_info_cache with network_info: [{"id": "810c9b00-6705-4e67-b4f6-f711f07f2037", "address": "fa:16:3e:a7:41:91", "network": {"id": "3341135c-a39c-4c4c-bc65-d60626d959cd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-815900522-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abb4ff37a7154d7da41e936b8358f93d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "821e0800-fabf-48d0-87b4-db5a1eddce93", "external-id": "nsx-vlan-transportzone-97", "segmentation_id": 97, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap810c9b00-67", "ovs_interfaceid": "810c9b00-6705-4e67-b4f6-f711f07f2037", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.211309] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239460, 'name': CreateVM_Task, 'duration_secs': 0.524975} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.212030] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.213097] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.213314] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.213797] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.214092] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbdd1609-007d-4d6a-b56b-4462e26619b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.219987] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 836.219987] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523a9f41-b084-017b-46b8-7d6d99ba0a0c" [ 836.219987] env[62552]: _type = "Task" [ 836.219987] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.232728] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523a9f41-b084-017b-46b8-7d6d99ba0a0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.312343] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239463, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.319105] env[62552]: DEBUG oslo_vmware.api [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239464, 'name': ResetVM_Task, 'duration_secs': 0.095764} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.319105] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Did hard reboot of VM {{(pid=62552) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 836.319105] env[62552]: DEBUG nova.compute.manager [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 836.319975] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acea99aa-1f49-42a4-a1c8-cc63b7298069 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.430412] env[62552]: DEBUG oslo_vmware.api [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239466, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326961} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.430864] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.431038] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.431255] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.464747] env[62552]: INFO nova.scheduler.client.report [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted allocations for instance cdee22a9-4327-47af-ab14-8403a06b2802 [ 836.497503] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.498977] env[62552]: DEBUG nova.compute.manager [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Instance network_info: |[{"id": "f749d931-7a86-41ca-a551-4f8d2013603a", "address": "fa:16:3e:00:02:b8", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf749d931-7a", "ovs_interfaceid": "f749d931-7a86-41ca-a551-4f8d2013603a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 836.499887] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:02:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e41c97-4d75-4041-ae71-321e7e9d480b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f749d931-7a86-41ca-a551-4f8d2013603a', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.508729] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Creating folder: Project (3daa5253a0cc4538903b7cabc2f98fa2). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.509906] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08eb9bea-e890-4be6-a6f6-2b13ed3d387c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.527328] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Created folder: Project (3daa5253a0cc4538903b7cabc2f98fa2) in parent group-v267339. [ 836.527652] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Creating folder: Instances. Parent ref: group-v267419. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.529258] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f04920c2-43c9-44f4-9866-982e782098b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.531927] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f640c9b3-12a4-48bf-9253-4c573072f3b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.542410] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7985b5c2-aaa2-437f-834c-f90b7cc3fedc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.547629] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Created folder: Instances in parent group-v267419. [ 836.547905] env[62552]: DEBUG oslo.service.loopingcall [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.548590] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.548817] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b6a29ea-cdd7-4e6d-9fd1-812d8ac494a1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.598586] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606da88d-3eab-433b-b73a-a20f035eb772 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.602484] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.602484] env[62552]: value = "task-1239469" [ 836.602484] env[62552]: _type = "Task" [ 836.602484] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.612963] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239469, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.614513] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d51fde4-0866-4e85-a1e0-14b98313728b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.632618] env[62552]: DEBUG nova.compute.provider_tree [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 836.685351] env[62552]: DEBUG oslo_concurrency.lockutils [req-256efa38-8e09-4f3c-8a7f-41d2131a6536 req-4fc3fee6-610c-4788-853f-fe5d58fdf45a service nova] Releasing lock "refresh_cache-c52efdeb-9efc-4ba1-b9af-4dd0a925517f" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.693644] env[62552]: INFO nova.compute.manager [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Took 0.58 seconds to detach 1 volumes for instance. [ 836.734028] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523a9f41-b084-017b-46b8-7d6d99ba0a0c, 'name': SearchDatastore_Task, 'duration_secs': 0.017309} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.734028] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.734028] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.734028] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.734028] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.734028] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.734028] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81870c0e-cdf4-415c-bea2-bcc60008c860 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.745684] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.745878] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.746894] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-037a3264-2a42-41bf-8a17-dee45d31bdde {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.752720] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 836.752720] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52260752-c4da-028a-cf30-60cebb4c2210" [ 836.752720] env[62552]: _type = "Task" [ 836.752720] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.766666] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52260752-c4da-028a-cf30-60cebb4c2210, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.806561] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239463, 'name': ReconfigVM_Task, 'duration_secs': 0.648361} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.806974] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 474dfa9e-d003-478f-b48f-09e0c0452ffd/474dfa9e-d003-478f-b48f-09e0c0452ffd.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.808247] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f5046cd-ee83-4cc6-a940-5e9086e6893f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.816997] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 836.816997] env[62552]: value = "task-1239470" [ 836.816997] env[62552]: _type = "Task" [ 836.816997] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.831496] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239470, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.837015] env[62552]: DEBUG oslo_concurrency.lockutils [None req-824a30c3-21c4-4818-8d1c-5bbe76093bf6 tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.710s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.970346] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.113493] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239469, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.163275] env[62552]: DEBUG nova.objects.instance [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'flavor' on Instance uuid 0418260a-aa27-4955-ab15-b180ec04f0b3 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.173909] env[62552]: DEBUG nova.scheduler.client.report [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 84 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 837.174198] env[62552]: DEBUG nova.compute.provider_tree [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 84 to 85 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 837.174379] env[62552]: DEBUG nova.compute.provider_tree [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 837.200894] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.264481] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52260752-c4da-028a-cf30-60cebb4c2210, 'name': SearchDatastore_Task, 'duration_secs': 0.031994} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.264637] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c2e4fc6-a024-4074-8154-681e943fa30d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.270755] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 837.270755] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5292d856-e090-1aa4-699e-9e401f449117" [ 837.270755] env[62552]: _type = "Task" [ 837.270755] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.280969] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5292d856-e090-1aa4-699e-9e401f449117, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.302416] env[62552]: DEBUG nova.compute.manager [req-a4eb9d01-b66a-4fe1-9f0a-891a67771c27 req-52541eb7-da6c-4350-8f87-e7b018b3e653 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Received event network-changed-f749d931-7a86-41ca-a551-4f8d2013603a {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 837.306479] env[62552]: DEBUG nova.compute.manager [req-a4eb9d01-b66a-4fe1-9f0a-891a67771c27 req-52541eb7-da6c-4350-8f87-e7b018b3e653 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Refreshing instance network info cache due to event network-changed-f749d931-7a86-41ca-a551-4f8d2013603a. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 837.306479] env[62552]: DEBUG oslo_concurrency.lockutils [req-a4eb9d01-b66a-4fe1-9f0a-891a67771c27 req-52541eb7-da6c-4350-8f87-e7b018b3e653 service nova] Acquiring lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.306479] env[62552]: DEBUG oslo_concurrency.lockutils [req-a4eb9d01-b66a-4fe1-9f0a-891a67771c27 req-52541eb7-da6c-4350-8f87-e7b018b3e653 service nova] Acquired lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.306479] env[62552]: DEBUG nova.network.neutron [req-a4eb9d01-b66a-4fe1-9f0a-891a67771c27 req-52541eb7-da6c-4350-8f87-e7b018b3e653 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Refreshing network info cache for port f749d931-7a86-41ca-a551-4f8d2013603a {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.330276] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239470, 'name': Rename_Task, 'duration_secs': 0.205108} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.330499] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.330787] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7827731a-f69f-45fa-8587-068544a5302c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.337382] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 837.337382] env[62552]: value = "task-1239471" [ 837.337382] env[62552]: _type = "Task" [ 837.337382] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.346080] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239471, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.613229] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239469, 'name': CreateVM_Task, 'duration_secs': 0.517425} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.613544] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.614096] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.614275] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.614610] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.614864] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14528234-02b6-46cd-bf40-e5b8f6dda94d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.619317] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 837.619317] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52eaeeb7-9847-b341-b281-9d298555fe5a" [ 837.619317] env[62552]: _type = "Task" [ 837.619317] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.627213] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52eaeeb7-9847-b341-b281-9d298555fe5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.668763] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4162cfe4-1380-4385-810d-c34b00b295ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.352s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.683235] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.854s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.683936] env[62552]: DEBUG nova.compute.manager [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 837.687720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.746s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.691025] env[62552]: INFO nova.compute.claims [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.784505] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5292d856-e090-1aa4-699e-9e401f449117, 'name': SearchDatastore_Task, 'duration_secs': 0.02579} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.785022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.785444] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] c52efdeb-9efc-4ba1-b9af-4dd0a925517f/c52efdeb-9efc-4ba1-b9af-4dd0a925517f.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.785871] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6f7feaa-7e90-478b-b56e-cb0ab465dcb8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.793147] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 837.793147] env[62552]: value = "task-1239472" [ 837.793147] env[62552]: _type = "Task" [ 837.793147] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.804075] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239472, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.849339] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239471, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.130014] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52eaeeb7-9847-b341-b281-9d298555fe5a, 'name': SearchDatastore_Task, 'duration_secs': 0.01155} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.130355] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.130652] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.131148] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.131331] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.131529] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.131848] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2a0f3c0-a67b-4579-b50a-2444214363e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.139899] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.140112] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.140887] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4b5db81-4b0d-492f-97e3-32e281a6dc95 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.146227] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 838.146227] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5281500f-c31c-3087-ab98-bdf2e58d8f7e" [ 838.146227] env[62552]: _type = "Task" [ 838.146227] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.154411] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5281500f-c31c-3087-ab98-bdf2e58d8f7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.195048] env[62552]: DEBUG nova.compute.utils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.202024] env[62552]: DEBUG nova.compute.manager [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 838.202024] env[62552]: DEBUG nova.network.neutron [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 838.303840] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239472, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.331438] env[62552]: DEBUG nova.policy [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd12de74ca7e43cbbce54ef31bb7b274', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aaedfcc2bfaf4fa4835cce89ee5abe8f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.348600] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239471, 'name': PowerOnVM_Task, 'duration_secs': 0.622234} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.348903] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.349155] env[62552]: INFO nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Took 9.77 seconds to spawn the instance on the hypervisor. [ 838.349433] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 838.350297] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2a8bc6-983e-4dde-83f1-ff646fe86549 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.511991] env[62552]: DEBUG nova.network.neutron [req-a4eb9d01-b66a-4fe1-9f0a-891a67771c27 req-52541eb7-da6c-4350-8f87-e7b018b3e653 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updated VIF entry in instance network info cache for port f749d931-7a86-41ca-a551-4f8d2013603a. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.511991] env[62552]: DEBUG nova.network.neutron [req-a4eb9d01-b66a-4fe1-9f0a-891a67771c27 req-52541eb7-da6c-4350-8f87-e7b018b3e653 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance_info_cache with network_info: [{"id": "f749d931-7a86-41ca-a551-4f8d2013603a", "address": "fa:16:3e:00:02:b8", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf749d931-7a", "ovs_interfaceid": "f749d931-7a86-41ca-a551-4f8d2013603a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.567100] env[62552]: DEBUG nova.compute.manager [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received event network-changed-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 838.568084] env[62552]: DEBUG nova.compute.manager [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Refreshing instance network info cache due to event network-changed-929d55b0-5a35-4d2d-a172-15d79a215977. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 838.568495] env[62552]: DEBUG oslo_concurrency.lockutils [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] Acquiring lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.568795] env[62552]: DEBUG oslo_concurrency.lockutils [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] Acquired lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.569091] env[62552]: DEBUG nova.network.neutron [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Refreshing network info cache for port 929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.657802] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5281500f-c31c-3087-ab98-bdf2e58d8f7e, 'name': SearchDatastore_Task, 'duration_secs': 0.010902} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.658847] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9def77c-fe2a-4c62-9c02-fb0af08a0cb0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.664774] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 838.664774] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527851ff-5cf1-3012-7586-7ab840a9bb36" [ 838.664774] env[62552]: _type = "Task" [ 838.664774] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.672524] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527851ff-5cf1-3012-7586-7ab840a9bb36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.702382] env[62552]: DEBUG nova.compute.manager [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 838.814908] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239472, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.776094} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.820486] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] c52efdeb-9efc-4ba1-b9af-4dd0a925517f/c52efdeb-9efc-4ba1-b9af-4dd0a925517f.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.820486] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.820737] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2e93148-b72d-401b-b8a8-51274f78d0f5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.828915] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 838.828915] env[62552]: value = "task-1239473" [ 838.828915] env[62552]: _type = "Task" [ 838.828915] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.842154] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239473, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.872185] env[62552]: INFO nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Took 36.54 seconds to build instance. [ 839.012784] env[62552]: DEBUG nova.network.neutron [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Successfully created port: aee2bcfd-d515-46fa-8574-2044d313840b {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.014884] env[62552]: DEBUG oslo_concurrency.lockutils [req-a4eb9d01-b66a-4fe1-9f0a-891a67771c27 req-52541eb7-da6c-4350-8f87-e7b018b3e653 service nova] Releasing lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.014989] env[62552]: DEBUG nova.compute.manager [req-a4eb9d01-b66a-4fe1-9f0a-891a67771c27 req-52541eb7-da6c-4350-8f87-e7b018b3e653 service nova] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Received event network-vif-deleted-3d60a1e0-8a29-42e5-a18e-f2a265bdb330 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 839.176964] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527851ff-5cf1-3012-7586-7ab840a9bb36, 'name': SearchDatastore_Task, 'duration_secs': 0.063292} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.177203] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.177445] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 22357d4e-9771-477c-9fc3-fe3d76f6e902/22357d4e-9771-477c-9fc3-fe3d76f6e902.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.177766] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-18f05909-0778-4607-aeed-814ad9dc9dec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.181071] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c87d13e-b43b-46b6-b5aa-0e3500d212bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.189670] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce04f09-118a-4469-8f30-fd38994517fc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.193410] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 839.193410] env[62552]: value = "task-1239474" [ 839.193410] env[62552]: _type = "Task" [ 839.193410] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.226388] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0091ce1-3a6d-4637-8c39-47a92d3b6a06 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.232708] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239474, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.240017] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c461bd-68e7-4a38-857e-f44bb9407624 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.256680] env[62552]: DEBUG nova.compute.provider_tree [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.344832] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239473, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090634} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.344832] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.346056] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22bc4d4-0a00-4a06-81ee-cd5c34a20ae3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.368612] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] c52efdeb-9efc-4ba1-b9af-4dd0a925517f/c52efdeb-9efc-4ba1-b9af-4dd0a925517f.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.371356] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95d53a10-e30e-4c6d-9bab-85a79f19f645 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.385869] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "474dfa9e-d003-478f-b48f-09e0c0452ffd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.895s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.393851] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 839.393851] env[62552]: value = "task-1239475" [ 839.393851] env[62552]: _type = "Task" [ 839.393851] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.404447] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239475, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.584573] env[62552]: DEBUG nova.network.neutron [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updated VIF entry in instance network info cache for port 929d55b0-5a35-4d2d-a172-15d79a215977. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.584977] env[62552]: DEBUG nova.network.neutron [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating instance_info_cache with network_info: [{"id": "929d55b0-5a35-4d2d-a172-15d79a215977", "address": "fa:16:3e:62:4d:9b", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap929d55b0-5a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.704129] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239474, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.731881] env[62552]: DEBUG nova.compute.manager [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 839.762309] env[62552]: DEBUG nova.scheduler.client.report [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 839.769630] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.770186] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.770186] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.770318] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.770462] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.770628] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.770844] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.771017] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.771219] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.771364] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.771532] env[62552]: DEBUG nova.virt.hardware [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.772515] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab26a5e0-0c3c-4b51-8531-7f2677bcc3b6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.786563] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d42dec-0921-40eb-842f-125c56c759ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.897841] env[62552]: DEBUG nova.compute.manager [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 839.904862] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239475, 'name': ReconfigVM_Task, 'duration_secs': 0.486577} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.904862] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Reconfigured VM instance instance-0000003e to attach disk [datastore2] c52efdeb-9efc-4ba1-b9af-4dd0a925517f/c52efdeb-9efc-4ba1-b9af-4dd0a925517f.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.905246] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65193e71-9aff-44f7-befe-02b2eb086128 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.912403] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 839.912403] env[62552]: value = "task-1239476" [ 839.912403] env[62552]: _type = "Task" [ 839.912403] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.925955] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239476, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.088471] env[62552]: DEBUG oslo_concurrency.lockutils [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] Releasing lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.088788] env[62552]: DEBUG nova.compute.manager [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Received event network-changed-43bb3381-583a-42f4-bcd9-3db694582a1d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 840.088977] env[62552]: DEBUG nova.compute.manager [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Refreshing instance network info cache due to event network-changed-43bb3381-583a-42f4-bcd9-3db694582a1d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 840.089219] env[62552]: DEBUG oslo_concurrency.lockutils [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] Acquiring lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.089370] env[62552]: DEBUG oslo_concurrency.lockutils [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] Acquired lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.089540] env[62552]: DEBUG nova.network.neutron [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Refreshing network info cache for port 43bb3381-583a-42f4-bcd9-3db694582a1d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.207458] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239474, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.807868} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.207458] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 22357d4e-9771-477c-9fc3-fe3d76f6e902/22357d4e-9771-477c-9fc3-fe3d76f6e902.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.207458] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.207458] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d022f284-19da-4a39-a7dd-8363b21f16aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.212602] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 840.212602] env[62552]: value = "task-1239477" [ 840.212602] env[62552]: _type = "Task" [ 840.212602] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.221208] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239477, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.267138] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.267826] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.279186] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.279288] env[62552]: DEBUG nova.compute.manager [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 840.283307] env[62552]: DEBUG oslo_concurrency.lockutils [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.338s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.412020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.424385] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239476, 'name': Rename_Task, 'duration_secs': 0.202008} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.424645] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.424789] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a987b83-6a68-4713-8071-4e891090f9ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.431233] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 840.431233] env[62552]: value = "task-1239478" [ 840.431233] env[62552]: _type = "Task" [ 840.431233] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.442767] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239478, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.484037] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.484201] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.484409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.484656] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.484910] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.488058] env[62552]: INFO nova.compute.manager [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Terminating instance [ 840.616282] env[62552]: DEBUG nova.compute.manager [req-c29731f9-3a32-4d6f-b516-800cfd22803e req-b272acf0-b1fe-43c5-b2e4-a18f80a96507 service nova] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Received event network-vif-plugged-aee2bcfd-d515-46fa-8574-2044d313840b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 840.616282] env[62552]: DEBUG oslo_concurrency.lockutils [req-c29731f9-3a32-4d6f-b516-800cfd22803e req-b272acf0-b1fe-43c5-b2e4-a18f80a96507 service nova] Acquiring lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.616282] env[62552]: DEBUG oslo_concurrency.lockutils [req-c29731f9-3a32-4d6f-b516-800cfd22803e req-b272acf0-b1fe-43c5-b2e4-a18f80a96507 service nova] Lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.616282] env[62552]: DEBUG oslo_concurrency.lockutils [req-c29731f9-3a32-4d6f-b516-800cfd22803e req-b272acf0-b1fe-43c5-b2e4-a18f80a96507 service nova] Lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.616282] env[62552]: DEBUG nova.compute.manager [req-c29731f9-3a32-4d6f-b516-800cfd22803e req-b272acf0-b1fe-43c5-b2e4-a18f80a96507 service nova] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] No waiting events found dispatching network-vif-plugged-aee2bcfd-d515-46fa-8574-2044d313840b {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 840.616465] env[62552]: WARNING nova.compute.manager [req-c29731f9-3a32-4d6f-b516-800cfd22803e req-b272acf0-b1fe-43c5-b2e4-a18f80a96507 service nova] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Received unexpected event network-vif-plugged-aee2bcfd-d515-46fa-8574-2044d313840b for instance with vm_state building and task_state spawning. [ 840.641276] env[62552]: DEBUG nova.compute.manager [req-3e7d981e-ccea-4c14-b01f-155a1ac233d4 req-958fa45e-6689-41a4-831b-d4b59a9f2ea5 service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Received event network-changed-61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 840.641276] env[62552]: DEBUG nova.compute.manager [req-3e7d981e-ccea-4c14-b01f-155a1ac233d4 req-958fa45e-6689-41a4-831b-d4b59a9f2ea5 service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Refreshing instance network info cache due to event network-changed-61602401-8f5b-4b30-8d2d-b326bb568984. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 840.641276] env[62552]: DEBUG oslo_concurrency.lockutils [req-3e7d981e-ccea-4c14-b01f-155a1ac233d4 req-958fa45e-6689-41a4-831b-d4b59a9f2ea5 service nova] Acquiring lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.641276] env[62552]: DEBUG oslo_concurrency.lockutils [req-3e7d981e-ccea-4c14-b01f-155a1ac233d4 req-958fa45e-6689-41a4-831b-d4b59a9f2ea5 service nova] Acquired lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.641276] env[62552]: DEBUG nova.network.neutron [req-3e7d981e-ccea-4c14-b01f-155a1ac233d4 req-958fa45e-6689-41a4-831b-d4b59a9f2ea5 service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Refreshing network info cache for port 61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.728872] env[62552]: DEBUG nova.objects.instance [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lazy-loading 'flavor' on Instance uuid 51bbcc4e-8251-4b38-9d36-8aea8fc7705d {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.734020] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239477, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.139043} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.734020] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.734020] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bbf51e-a999-48ae-a535-9eff05153bb4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.743754] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700f5f8a-1861-4635-b6d7-ea0db82cf47f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.773983] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 22357d4e-9771-477c-9fc3-fe3d76f6e902/22357d4e-9771-477c-9fc3-fe3d76f6e902.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.779492] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54ad57a7-c571-4f96-a958-4384d7a7d75c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.799981] env[62552]: DEBUG nova.compute.utils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.803382] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbaa4cae-6aad-4262-9566-7fdf44a13b94 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.810048] env[62552]: DEBUG nova.network.neutron [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Successfully updated port: aee2bcfd-d515-46fa-8574-2044d313840b {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.811508] env[62552]: DEBUG nova.compute.manager [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 840.811689] env[62552]: DEBUG nova.network.neutron [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.868645] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 840.868645] env[62552]: value = "task-1239479" [ 840.868645] env[62552]: _type = "Task" [ 840.868645] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.885298] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ca54b4-0025-4687-84d9-10e1b6f1ca32 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.914756] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320a441d-3804-4afa-99b1-d89449b2948f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.924554] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239479, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.944098] env[62552]: DEBUG nova.compute.provider_tree [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.953591] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239478, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.992026] env[62552]: DEBUG nova.compute.manager [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 840.992112] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.992934] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4579e9-15a7-4fee-9b26-45a30cfdea76 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.002230] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 841.002230] env[62552]: DEBUG nova.policy [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd12de74ca7e43cbbce54ef31bb7b274', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aaedfcc2bfaf4fa4835cce89ee5abe8f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.003532] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17fced82-4039-4eab-a931-6f5946521069 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.009682] env[62552]: DEBUG oslo_vmware.api [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 841.009682] env[62552]: value = "task-1239480" [ 841.009682] env[62552]: _type = "Task" [ 841.009682] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.018352] env[62552]: DEBUG oslo_vmware.api [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.079170] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "cdee22a9-4327-47af-ab14-8403a06b2802" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.086556] env[62552]: DEBUG nova.network.neutron [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updated VIF entry in instance network info cache for port 43bb3381-583a-42f4-bcd9-3db694582a1d. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 841.086962] env[62552]: DEBUG nova.network.neutron [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.236484] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.237229] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquired lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.312150] env[62552]: DEBUG nova.compute.manager [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 841.320495] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "refresh_cache-d8891025-5bdf-4dc9-a2b0-c86f94582ac6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.320495] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquired lock "refresh_cache-d8891025-5bdf-4dc9-a2b0-c86f94582ac6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.320495] env[62552]: DEBUG nova.network.neutron [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.391803] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239479, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.453510] env[62552]: DEBUG nova.scheduler.client.report [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 841.456962] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239478, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.525362] env[62552]: DEBUG oslo_vmware.api [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239480, 'name': PowerOffVM_Task, 'duration_secs': 0.297019} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.525362] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 841.525583] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 841.526035] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-559c4a87-69f0-48f7-953c-558ec1274251 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.544379] env[62552]: DEBUG nova.network.neutron [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Successfully created port: b77f3748-ccd8-4287-9ed2-55c69d3c783f {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.590841] env[62552]: DEBUG oslo_concurrency.lockutils [req-5a796ffd-a898-4aae-9901-e68e636ab9c2 req-a488f206-f099-48ba-b32b-71f1f1a2d531 service nova] Releasing lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.600283] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.600515] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.600762] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Deleting the datastore file [datastore1] df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.601061] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0196a2e9-cedf-4acb-b2f6-c02419cc95ca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.608586] env[62552]: DEBUG oslo_vmware.api [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for the task: (returnval){ [ 841.608586] env[62552]: value = "task-1239482" [ 841.608586] env[62552]: _type = "Task" [ 841.608586] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.620300] env[62552]: DEBUG oslo_vmware.api [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.829075] env[62552]: DEBUG nova.network.neutron [req-3e7d981e-ccea-4c14-b01f-155a1ac233d4 req-958fa45e-6689-41a4-831b-d4b59a9f2ea5 service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Updated VIF entry in instance network info cache for port 61602401-8f5b-4b30-8d2d-b326bb568984. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 841.829425] env[62552]: DEBUG nova.network.neutron [req-3e7d981e-ccea-4c14-b01f-155a1ac233d4 req-958fa45e-6689-41a4-831b-d4b59a9f2ea5 service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Updating instance_info_cache with network_info: [{"id": "61602401-8f5b-4b30-8d2d-b326bb568984", "address": "fa:16:3e:06:29:86", "network": {"id": "3391f646-fd85-4a2a-b4a1-5d92620ed41b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1518992046-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd93b575f2e04fc9b93319545fa03708", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61602401-8f", "ovs_interfaceid": "61602401-8f5b-4b30-8d2d-b326bb568984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.894163] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239479, 'name': ReconfigVM_Task, 'duration_secs': 0.708922} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.894527] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 22357d4e-9771-477c-9fc3-fe3d76f6e902/22357d4e-9771-477c-9fc3-fe3d76f6e902.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.895394] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d14268bd-8418-4f68-96a2-1897d77c2f39 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.902118] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 841.902118] env[62552]: value = "task-1239483" [ 841.902118] env[62552]: _type = "Task" [ 841.902118] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.906092] env[62552]: DEBUG nova.network.neutron [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.915180] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239483, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.951823] env[62552]: DEBUG oslo_vmware.api [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239478, 'name': PowerOnVM_Task, 'duration_secs': 1.104953} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.952191] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.952451] env[62552]: INFO nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Took 10.74 seconds to spawn the instance on the hypervisor. [ 841.952680] env[62552]: DEBUG nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 841.953517] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a7b4dc-b072-4f69-9571-dd7bdb416710 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.964053] env[62552]: DEBUG oslo_concurrency.lockutils [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.681s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.964531] env[62552]: INFO nova.compute.manager [None req-40cab95e-b279-4827-b219-a09677c12bfa tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Successfully reverted task state from rebuilding on failure for instance. [ 841.971752] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.938s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.973293] env[62552]: INFO nova.compute.claims [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.118651] env[62552]: DEBUG oslo_vmware.api [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Task: {'id': task-1239482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253552} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.118937] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.119179] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 842.119363] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 842.119566] env[62552]: INFO nova.compute.manager [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Took 1.13 seconds to destroy the instance on the hypervisor. [ 842.119946] env[62552]: DEBUG oslo.service.loopingcall [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.120023] env[62552]: DEBUG nova.compute.manager [-] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 842.120082] env[62552]: DEBUG nova.network.neutron [-] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 842.288227] env[62552]: DEBUG nova.network.neutron [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Updating instance_info_cache with network_info: [{"id": "aee2bcfd-d515-46fa-8574-2044d313840b", "address": "fa:16:3e:88:45:6f", "network": {"id": "bb6a7df6-f61d-4964-be8b-1c2376e2716d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.57", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "04a4549325ec47ddb8b99e2a28b02183", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a0d5af-5be9-477a-837c-58ef55c717f4", "external-id": "nsx-vlan-transportzone-598", "segmentation_id": 598, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaee2bcfd-d5", "ovs_interfaceid": "aee2bcfd-d515-46fa-8574-2044d313840b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.297827] env[62552]: DEBUG nova.network.neutron [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.328229] env[62552]: DEBUG nova.compute.manager [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 842.333059] env[62552]: DEBUG oslo_concurrency.lockutils [req-3e7d981e-ccea-4c14-b01f-155a1ac233d4 req-958fa45e-6689-41a4-831b-d4b59a9f2ea5 service nova] Releasing lock "refresh_cache-df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.362099] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.362099] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.362384] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.362384] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.362544] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.362672] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.362872] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.363074] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.363246] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.363416] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.367020] env[62552]: DEBUG nova.virt.hardware [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.367020] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9341d595-5980-4748-90d9-5511c0fa1355 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.372612] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c961b0a-4d03-4b25-a045-18783778ff92 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.412816] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239483, 'name': Rename_Task, 'duration_secs': 0.136819} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.413181] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.413450] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45e98bdc-62e6-4772-9417-a0a453df49de {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.426574] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 842.426574] env[62552]: value = "task-1239484" [ 842.426574] env[62552]: _type = "Task" [ 842.426574] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.440446] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239484, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.484433] env[62552]: INFO nova.compute.manager [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Took 39.55 seconds to build instance. [ 842.801022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Releasing lock "refresh_cache-d8891025-5bdf-4dc9-a2b0-c86f94582ac6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.801022] env[62552]: DEBUG nova.compute.manager [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Instance network_info: |[{"id": "aee2bcfd-d515-46fa-8574-2044d313840b", "address": "fa:16:3e:88:45:6f", "network": {"id": "bb6a7df6-f61d-4964-be8b-1c2376e2716d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.57", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "04a4549325ec47ddb8b99e2a28b02183", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a0d5af-5be9-477a-837c-58ef55c717f4", "external-id": "nsx-vlan-transportzone-598", "segmentation_id": 598, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaee2bcfd-d5", "ovs_interfaceid": "aee2bcfd-d515-46fa-8574-2044d313840b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 842.801022] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:45:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7a0d5af-5be9-477a-837c-58ef55c717f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aee2bcfd-d515-46fa-8574-2044d313840b', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.810091] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Creating folder: Project (aaedfcc2bfaf4fa4835cce89ee5abe8f). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.813547] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6dbb25ef-1dfc-40c8-94e7-9b20259807e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.829217] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Created folder: Project (aaedfcc2bfaf4fa4835cce89ee5abe8f) in parent group-v267339. [ 842.829485] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Creating folder: Instances. Parent ref: group-v267422. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.829726] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c51e223-e250-4e79-89cd-67fe5e5127e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.842999] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Created folder: Instances in parent group-v267422. [ 842.843234] env[62552]: DEBUG oslo.service.loopingcall [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.843440] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.843654] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ff208f8-8a7b-4bfe-b7c7-3d07f21dd1b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.874151] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.874151] env[62552]: value = "task-1239487" [ 842.874151] env[62552]: _type = "Task" [ 842.874151] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.884146] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239487, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.939516] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239484, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.992956] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9f76224d-9542-4dc5-ade5-34f40b714647 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.473s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.082176] env[62552]: DEBUG nova.network.neutron [-] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.303686] env[62552]: DEBUG nova.compute.manager [req-e4341277-4af8-40db-b2ab-3744552998e2 req-151eb4a2-38e9-4364-81c9-f8350fdb011d service nova] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Received event network-changed-aee2bcfd-d515-46fa-8574-2044d313840b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 843.303868] env[62552]: DEBUG nova.compute.manager [req-e4341277-4af8-40db-b2ab-3744552998e2 req-151eb4a2-38e9-4364-81c9-f8350fdb011d service nova] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Refreshing instance network info cache due to event network-changed-aee2bcfd-d515-46fa-8574-2044d313840b. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 843.304131] env[62552]: DEBUG oslo_concurrency.lockutils [req-e4341277-4af8-40db-b2ab-3744552998e2 req-151eb4a2-38e9-4364-81c9-f8350fdb011d service nova] Acquiring lock "refresh_cache-d8891025-5bdf-4dc9-a2b0-c86f94582ac6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.304288] env[62552]: DEBUG oslo_concurrency.lockutils [req-e4341277-4af8-40db-b2ab-3744552998e2 req-151eb4a2-38e9-4364-81c9-f8350fdb011d service nova] Acquired lock "refresh_cache-d8891025-5bdf-4dc9-a2b0-c86f94582ac6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.304479] env[62552]: DEBUG nova.network.neutron [req-e4341277-4af8-40db-b2ab-3744552998e2 req-151eb4a2-38e9-4364-81c9-f8350fdb011d service nova] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Refreshing network info cache for port aee2bcfd-d515-46fa-8574-2044d313840b {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 843.334085] env[62552]: DEBUG nova.network.neutron [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Successfully updated port: b77f3748-ccd8-4287-9ed2-55c69d3c783f {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.336598] env[62552]: DEBUG nova.compute.manager [req-714d0ace-7929-47af-8232-8f9cb24a613f req-adfd04a4-5cf3-4923-b645-5d90c649b2d3 service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Received event network-changed-e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 843.336742] env[62552]: DEBUG nova.compute.manager [req-714d0ace-7929-47af-8232-8f9cb24a613f req-adfd04a4-5cf3-4923-b645-5d90c649b2d3 service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Refreshing instance network info cache due to event network-changed-e3439c52-55fc-4ea8-ae30-607c154031cb. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 843.337375] env[62552]: DEBUG oslo_concurrency.lockutils [req-714d0ace-7929-47af-8232-8f9cb24a613f req-adfd04a4-5cf3-4923-b645-5d90c649b2d3 service nova] Acquiring lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.385128] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239487, 'name': CreateVM_Task, 'duration_secs': 0.489089} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.385315] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.386757] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.386936] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.387303] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.387569] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d56faa87-84d3-43de-a5a2-20c10548e7b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.393487] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 843.393487] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523fc869-fc78-8bdb-2fb3-b7bbcb3b9f69" [ 843.393487] env[62552]: _type = "Task" [ 843.393487] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.407737] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523fc869-fc78-8bdb-2fb3-b7bbcb3b9f69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.447741] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239484, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.478044] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8bb33f-dc41-411e-a794-a7c08cc4cc58 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.486375] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760cb757-6885-4385-b651-7174ee067923 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.525018] env[62552]: DEBUG nova.compute.manager [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 843.530938] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b10da3a-6b34-4290-8ebc-5422e2743913 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.543033] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a2296b-5009-45f2-88e7-68467081a710 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.557355] env[62552]: DEBUG nova.compute.provider_tree [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.559640] env[62552]: DEBUG nova.network.neutron [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updating instance_info_cache with network_info: [{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.565308] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquiring lock "0080a6b3-cac0-44f3-a037-3f94f1daa275" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.566044] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "0080a6b3-cac0-44f3-a037-3f94f1daa275" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.585278] env[62552]: INFO nova.compute.manager [-] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Took 1.47 seconds to deallocate network for instance. [ 843.839399] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "refresh_cache-4083ee43-ecea-4ea5-8923-42b348893824" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.839880] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquired lock "refresh_cache-4083ee43-ecea-4ea5-8923-42b348893824" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.839880] env[62552]: DEBUG nova.network.neutron [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.909631] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523fc869-fc78-8bdb-2fb3-b7bbcb3b9f69, 'name': SearchDatastore_Task, 'duration_secs': 0.014211} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.909944] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.910203] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.910434] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.910619] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.910797] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.911078] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b13f5fc-7554-4a1a-8e25-b3f3fc2996f0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.920991] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.921214] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.921951] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6158790a-c43f-4649-839e-5635fcf445ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.928089] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 843.928089] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52be3f29-ae2e-316e-b3d6-81c4a1f1a17a" [ 843.928089] env[62552]: _type = "Task" [ 843.928089] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.939338] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52be3f29-ae2e-316e-b3d6-81c4a1f1a17a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.942721] env[62552]: DEBUG oslo_vmware.api [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239484, 'name': PowerOnVM_Task, 'duration_secs': 1.185568} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.942927] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.943115] env[62552]: INFO nova.compute.manager [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Took 10.15 seconds to spawn the instance on the hypervisor. [ 843.943319] env[62552]: DEBUG nova.compute.manager [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 843.944101] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96628d98-a8c1-4c24-afe0-638268917e44 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.057955] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.064046] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Releasing lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.067686] env[62552]: DEBUG nova.compute.manager [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Inject network info {{(pid=62552) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7562}} [ 844.067686] env[62552]: DEBUG nova.compute.manager [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] network_info to inject: |[{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7563}} [ 844.070501] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Reconfiguring VM instance to set the machine id {{(pid=62552) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 844.072062] env[62552]: DEBUG oslo_concurrency.lockutils [req-714d0ace-7929-47af-8232-8f9cb24a613f req-adfd04a4-5cf3-4923-b645-5d90c649b2d3 service nova] Acquired lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.072719] env[62552]: DEBUG nova.network.neutron [req-714d0ace-7929-47af-8232-8f9cb24a613f req-adfd04a4-5cf3-4923-b645-5d90c649b2d3 service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Refreshing network info cache for port e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.074089] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b222b6f5-8e87-431a-95d4-d2491c6c72a5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.092020] env[62552]: ERROR nova.scheduler.client.report [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [req-01031a08-ddfe-4e62-a2e8-ee159b538521] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-01031a08-ddfe-4e62-a2e8-ee159b538521"}]} [ 844.097090] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.097857] env[62552]: DEBUG oslo_vmware.api [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 844.097857] env[62552]: value = "task-1239488" [ 844.097857] env[62552]: _type = "Task" [ 844.097857] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.111241] env[62552]: DEBUG oslo_vmware.api [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239488, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.111241] env[62552]: DEBUG nova.scheduler.client.report [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 844.125381] env[62552]: DEBUG nova.scheduler.client.report [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 844.125616] env[62552]: DEBUG nova.compute.provider_tree [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 844.140651] env[62552]: DEBUG nova.scheduler.client.report [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 844.163330] env[62552]: DEBUG nova.scheduler.client.report [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 844.179433] env[62552]: DEBUG nova.network.neutron [req-e4341277-4af8-40db-b2ab-3744552998e2 req-151eb4a2-38e9-4364-81c9-f8350fdb011d service nova] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Updated VIF entry in instance network info cache for port aee2bcfd-d515-46fa-8574-2044d313840b. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 844.179706] env[62552]: DEBUG nova.network.neutron [req-e4341277-4af8-40db-b2ab-3744552998e2 req-151eb4a2-38e9-4364-81c9-f8350fdb011d service nova] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Updating instance_info_cache with network_info: [{"id": "aee2bcfd-d515-46fa-8574-2044d313840b", "address": "fa:16:3e:88:45:6f", "network": {"id": "bb6a7df6-f61d-4964-be8b-1c2376e2716d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.57", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "04a4549325ec47ddb8b99e2a28b02183", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a0d5af-5be9-477a-837c-58ef55c717f4", "external-id": "nsx-vlan-transportzone-598", "segmentation_id": 598, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaee2bcfd-d5", "ovs_interfaceid": "aee2bcfd-d515-46fa-8574-2044d313840b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.385596] env[62552]: DEBUG nova.network.neutron [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.442620] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52be3f29-ae2e-316e-b3d6-81c4a1f1a17a, 'name': SearchDatastore_Task, 'duration_secs': 0.023602} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.445812] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74803902-24a0-4c33-aba0-3ff7a8bd6c77 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.463067] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 844.463067] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525046b3-93ba-a39f-8ffa-5ef43c82532d" [ 844.463067] env[62552]: _type = "Task" [ 844.463067] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.471451] env[62552]: INFO nova.compute.manager [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Took 40.85 seconds to build instance. [ 844.478978] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525046b3-93ba-a39f-8ffa-5ef43c82532d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.590949] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf72ef89-495a-4e65-a1b1-6468b78785c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.597989] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb6be22-4b59-48c6-bef0-ace07f288e5a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.605015] env[62552]: DEBUG nova.network.neutron [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Updating instance_info_cache with network_info: [{"id": "b77f3748-ccd8-4287-9ed2-55c69d3c783f", "address": "fa:16:3e:d3:92:2f", "network": {"id": "bb6a7df6-f61d-4964-be8b-1c2376e2716d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.211", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "04a4549325ec47ddb8b99e2a28b02183", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a0d5af-5be9-477a-837c-58ef55c717f4", "external-id": "nsx-vlan-transportzone-598", "segmentation_id": 598, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb77f3748-cc", "ovs_interfaceid": "b77f3748-ccd8-4287-9ed2-55c69d3c783f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.631910] env[62552]: DEBUG oslo_vmware.api [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239488, 'name': ReconfigVM_Task, 'duration_secs': 0.166025} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.635493] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Releasing lock "refresh_cache-4083ee43-ecea-4ea5-8923-42b348893824" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.635773] env[62552]: DEBUG nova.compute.manager [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Instance network_info: |[{"id": "b77f3748-ccd8-4287-9ed2-55c69d3c783f", "address": "fa:16:3e:d3:92:2f", "network": {"id": "bb6a7df6-f61d-4964-be8b-1c2376e2716d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.211", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "04a4549325ec47ddb8b99e2a28b02183", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a0d5af-5be9-477a-837c-58ef55c717f4", "external-id": "nsx-vlan-transportzone-598", "segmentation_id": 598, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb77f3748-cc", "ovs_interfaceid": "b77f3748-ccd8-4287-9ed2-55c69d3c783f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 844.636072] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c61c6aa5-8b31-4850-b6a2-b0375c64293c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Reconfigured VM instance to set the machine id {{(pid=62552) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 844.639397] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47510dce-9eb8-4d89-937d-87e403685a62 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.642395] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:92:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7a0d5af-5be9-477a-837c-58ef55c717f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b77f3748-ccd8-4287-9ed2-55c69d3c783f', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.649767] env[62552]: DEBUG oslo.service.loopingcall [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.650708] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.651356] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b11f2ca-5b04-4e64-ae1f-7e067a58b669 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.668457] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "474dfa9e-d003-478f-b48f-09e0c0452ffd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.668746] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "474dfa9e-d003-478f-b48f-09e0c0452ffd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.668966] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "474dfa9e-d003-478f-b48f-09e0c0452ffd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.669224] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "474dfa9e-d003-478f-b48f-09e0c0452ffd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.669422] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "474dfa9e-d003-478f-b48f-09e0c0452ffd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.672731] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59666173-13be-4cc3-bd5e-8085fb821c2a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.676662] env[62552]: INFO nova.compute.manager [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Terminating instance [ 844.679243] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.679243] env[62552]: value = "task-1239489" [ 844.679243] env[62552]: _type = "Task" [ 844.679243] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.694841] env[62552]: DEBUG oslo_concurrency.lockutils [req-e4341277-4af8-40db-b2ab-3744552998e2 req-151eb4a2-38e9-4364-81c9-f8350fdb011d service nova] Releasing lock "refresh_cache-d8891025-5bdf-4dc9-a2b0-c86f94582ac6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.695447] env[62552]: DEBUG nova.compute.provider_tree [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 844.703578] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239489, 'name': CreateVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.748252] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.748856] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.748856] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.748974] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.749188] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.752572] env[62552]: INFO nova.compute.manager [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Terminating instance [ 844.905453] env[62552]: DEBUG nova.network.neutron [req-714d0ace-7929-47af-8232-8f9cb24a613f req-adfd04a4-5cf3-4923-b645-5d90c649b2d3 service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updated VIF entry in instance network info cache for port e3439c52-55fc-4ea8-ae30-607c154031cb. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 844.905845] env[62552]: DEBUG nova.network.neutron [req-714d0ace-7929-47af-8232-8f9cb24a613f req-adfd04a4-5cf3-4923-b645-5d90c649b2d3 service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updating instance_info_cache with network_info: [{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.910586] env[62552]: DEBUG nova.objects.instance [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lazy-loading 'flavor' on Instance uuid 51bbcc4e-8251-4b38-9d36-8aea8fc7705d {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.976635] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525046b3-93ba-a39f-8ffa-5ef43c82532d, 'name': SearchDatastore_Task, 'duration_secs': 0.033753} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.977036] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.977557] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] d8891025-5bdf-4dc9-a2b0-c86f94582ac6/d8891025-5bdf-4dc9-a2b0-c86f94582ac6.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.977557] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-530a5147-8f95-4147-991e-14515687dbd1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.980313] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ddd3fa22-3d07-400a-ad0c-c4d8d3e2c906 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.883s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.988087] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 844.988087] env[62552]: value = "task-1239490" [ 844.988087] env[62552]: _type = "Task" [ 844.988087] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.996861] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.181696] env[62552]: DEBUG nova.compute.manager [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 845.182079] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.187024] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e735fc1-7a12-4e16-b819-8e2819edfd5c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.194842] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239489, 'name': CreateVM_Task, 'duration_secs': 0.493604} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.196737] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.197044] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.197692] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.197856] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.198205] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.198435] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a85025df-043a-4d53-824a-32877fbda9de {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.199926] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adae2bff-2e40-4d18-9a6f-34a48989995f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.207654] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 845.207654] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b0273c-9efe-e03a-f52e-353344cff303" [ 845.207654] env[62552]: _type = "Task" [ 845.207654] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.209150] env[62552]: DEBUG oslo_vmware.api [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 845.209150] env[62552]: value = "task-1239491" [ 845.209150] env[62552]: _type = "Task" [ 845.209150] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.222160] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b0273c-9efe-e03a-f52e-353344cff303, 'name': SearchDatastore_Task} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.225780] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.226154] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.226465] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.226646] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.226847] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.227172] env[62552]: DEBUG oslo_vmware.api [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.227421] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5fc4cbbb-35b4-46c6-96fd-8049542eed7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.235937] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.236217] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.237626] env[62552]: DEBUG nova.scheduler.client.report [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 845.237913] env[62552]: DEBUG nova.compute.provider_tree [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 88 to 89 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 845.238174] env[62552]: DEBUG nova.compute.provider_tree [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 845.241949] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef0d13ee-36de-4467-be75-a5d592cefb10 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.247805] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 845.247805] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5237067b-1055-f22c-3f41-5ef30089410c" [ 845.247805] env[62552]: _type = "Task" [ 845.247805] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.256508] env[62552]: DEBUG nova.compute.manager [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 845.256784] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.257138] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5237067b-1055-f22c-3f41-5ef30089410c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.258669] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a0a5969-411c-4ee9-a781-11e1386058a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.266947] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.267247] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d6d470f-75ae-4e2e-b4b8-21475632db05 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.277851] env[62552]: DEBUG oslo_vmware.api [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 845.277851] env[62552]: value = "task-1239492" [ 845.277851] env[62552]: _type = "Task" [ 845.277851] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.286855] env[62552]: DEBUG oslo_vmware.api [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239492, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.365141] env[62552]: DEBUG nova.compute.manager [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Received event network-vif-plugged-b77f3748-ccd8-4287-9ed2-55c69d3c783f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 845.365451] env[62552]: DEBUG oslo_concurrency.lockutils [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] Acquiring lock "4083ee43-ecea-4ea5-8923-42b348893824-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.365703] env[62552]: DEBUG oslo_concurrency.lockutils [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] Lock "4083ee43-ecea-4ea5-8923-42b348893824-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.365931] env[62552]: DEBUG oslo_concurrency.lockutils [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] Lock "4083ee43-ecea-4ea5-8923-42b348893824-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.367025] env[62552]: DEBUG nova.compute.manager [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] No waiting events found dispatching network-vif-plugged-b77f3748-ccd8-4287-9ed2-55c69d3c783f {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 845.367025] env[62552]: WARNING nova.compute.manager [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Received unexpected event network-vif-plugged-b77f3748-ccd8-4287-9ed2-55c69d3c783f for instance with vm_state building and task_state spawning. [ 845.367025] env[62552]: DEBUG nova.compute.manager [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Received event network-changed-b77f3748-ccd8-4287-9ed2-55c69d3c783f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 845.367025] env[62552]: DEBUG nova.compute.manager [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Refreshing instance network info cache due to event network-changed-b77f3748-ccd8-4287-9ed2-55c69d3c783f. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 845.367025] env[62552]: DEBUG oslo_concurrency.lockutils [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] Acquiring lock "refresh_cache-4083ee43-ecea-4ea5-8923-42b348893824" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.367025] env[62552]: DEBUG oslo_concurrency.lockutils [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] Acquired lock "refresh_cache-4083ee43-ecea-4ea5-8923-42b348893824" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.367346] env[62552]: DEBUG nova.network.neutron [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Refreshing network info cache for port b77f3748-ccd8-4287-9ed2-55c69d3c783f {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 845.408855] env[62552]: DEBUG oslo_concurrency.lockutils [req-714d0ace-7929-47af-8232-8f9cb24a613f req-adfd04a4-5cf3-4923-b645-5d90c649b2d3 service nova] Releasing lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.409190] env[62552]: DEBUG nova.compute.manager [req-714d0ace-7929-47af-8232-8f9cb24a613f req-adfd04a4-5cf3-4923-b645-5d90c649b2d3 service nova] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Received event network-vif-deleted-61602401-8f5b-4b30-8d2d-b326bb568984 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 845.415779] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.415977] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquired lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.491430] env[62552]: DEBUG nova.compute.manager [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 845.501844] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239490, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.722258] env[62552]: DEBUG oslo_vmware.api [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239491, 'name': PowerOffVM_Task, 'duration_secs': 0.207764} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.722608] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.723471] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 845.723471] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-674ca768-c8e9-4bb9-b6a7-de6df4c2aba0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.750431] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.775s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.750431] env[62552]: DEBUG nova.compute.manager [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 845.751388] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.024s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.753454] env[62552]: INFO nova.compute.claims [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.768841] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5237067b-1055-f22c-3f41-5ef30089410c, 'name': SearchDatastore_Task, 'duration_secs': 0.057547} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.769494] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f507009-b640-4559-bd4f-4b1a0adff6c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.774358] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 845.774358] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524653e5-5c35-142b-2189-9165d11f2db3" [ 845.774358] env[62552]: _type = "Task" [ 845.774358] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.785214] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524653e5-5c35-142b-2189-9165d11f2db3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.792350] env[62552]: DEBUG oslo_vmware.api [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239492, 'name': PowerOffVM_Task, 'duration_secs': 0.257929} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.792833] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.793154] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 845.793402] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-58dfb69f-7aba-49f2-9333-50efd16d59ac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.801036] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 845.801263] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 845.801491] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Deleting the datastore file [datastore1] 474dfa9e-d003-478f-b48f-09e0c0452ffd {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.801801] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37cb9434-22b2-4421-a4b7-40bd4aa1e6b8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.808599] env[62552]: DEBUG oslo_vmware.api [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 845.808599] env[62552]: value = "task-1239495" [ 845.808599] env[62552]: _type = "Task" [ 845.808599] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.818159] env[62552]: DEBUG oslo_vmware.api [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239495, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.829484] env[62552]: DEBUG nova.network.neutron [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 845.866022] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 845.866022] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 845.866022] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Deleting the datastore file [datastore2] c52efdeb-9efc-4ba1-b9af-4dd0a925517f {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.866022] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d12c54b1-9d5c-401a-b3fa-4b2f30918698 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.873655] env[62552]: DEBUG oslo_vmware.api [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for the task: (returnval){ [ 845.873655] env[62552]: value = "task-1239496" [ 845.873655] env[62552]: _type = "Task" [ 845.873655] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.885014] env[62552]: DEBUG oslo_vmware.api [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239496, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.001565] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645599} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.004011] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] d8891025-5bdf-4dc9-a2b0-c86f94582ac6/d8891025-5bdf-4dc9-a2b0-c86f94582ac6.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.004245] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.004733] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-781ec742-1877-4495-95c9-dc5e6effb37f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.014033] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 846.014033] env[62552]: value = "task-1239497" [ 846.014033] env[62552]: _type = "Task" [ 846.014033] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.020807] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239497, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.022022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.266773] env[62552]: DEBUG nova.compute.utils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.268688] env[62552]: DEBUG nova.compute.manager [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 846.268856] env[62552]: DEBUG nova.network.neutron [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.284520] env[62552]: DEBUG nova.network.neutron [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Updated VIF entry in instance network info cache for port b77f3748-ccd8-4287-9ed2-55c69d3c783f. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 846.284913] env[62552]: DEBUG nova.network.neutron [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Updating instance_info_cache with network_info: [{"id": "b77f3748-ccd8-4287-9ed2-55c69d3c783f", "address": "fa:16:3e:d3:92:2f", "network": {"id": "bb6a7df6-f61d-4964-be8b-1c2376e2716d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.211", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "04a4549325ec47ddb8b99e2a28b02183", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a0d5af-5be9-477a-837c-58ef55c717f4", "external-id": "nsx-vlan-transportzone-598", "segmentation_id": 598, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb77f3748-cc", "ovs_interfaceid": "b77f3748-ccd8-4287-9ed2-55c69d3c783f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.294313] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524653e5-5c35-142b-2189-9165d11f2db3, 'name': SearchDatastore_Task, 'duration_secs': 0.032869} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.294619] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.295033] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 4083ee43-ecea-4ea5-8923-42b348893824/4083ee43-ecea-4ea5-8923-42b348893824.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.295346] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6b3538c-3e8a-4d6c-9482-f20505c88198 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.304289] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 846.304289] env[62552]: value = "task-1239498" [ 846.304289] env[62552]: _type = "Task" [ 846.304289] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.315740] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.320828] env[62552]: DEBUG oslo_vmware.api [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239495, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.361714} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.321111] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.321302] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 846.322239] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 846.322239] env[62552]: INFO nova.compute.manager [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Took 1.14 seconds to destroy the instance on the hypervisor. [ 846.322239] env[62552]: DEBUG oslo.service.loopingcall [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.322239] env[62552]: DEBUG nova.compute.manager [-] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 846.322239] env[62552]: DEBUG nova.network.neutron [-] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 846.380807] env[62552]: DEBUG nova.policy [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6515846d33e44bca942767f9e1d650b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2f57f72de0c4adc9754458f4e7040b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.387534] env[62552]: DEBUG oslo_vmware.api [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Task: {'id': task-1239496, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.351202} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.387773] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.388019] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 846.388237] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 846.388432] env[62552]: INFO nova.compute.manager [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 846.388678] env[62552]: DEBUG oslo.service.loopingcall [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.388875] env[62552]: DEBUG nova.compute.manager [-] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 846.388967] env[62552]: DEBUG nova.network.neutron [-] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 846.524279] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239497, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070388} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.524615] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.525417] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a91d0f0-1d3c-4fe0-bb8c-24d307749b78 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.549628] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] d8891025-5bdf-4dc9-a2b0-c86f94582ac6/d8891025-5bdf-4dc9-a2b0-c86f94582ac6.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.551810] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b39d5ef-c484-4031-96a8-ce2101e3611b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.573685] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 846.573685] env[62552]: value = "task-1239499" [ 846.573685] env[62552]: _type = "Task" [ 846.573685] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.595022] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.775575] env[62552]: DEBUG nova.compute.manager [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 846.787559] env[62552]: DEBUG oslo_concurrency.lockutils [req-f0de54d5-d1b3-4f02-8a70-5cab35d81674 req-7992fcfc-a001-4391-b4e3-23375887392e service nova] Releasing lock "refresh_cache-4083ee43-ecea-4ea5-8923-42b348893824" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.820326] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239498, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.048641] env[62552]: DEBUG nova.network.neutron [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updating instance_info_cache with network_info: [{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.087126] env[62552]: DEBUG nova.network.neutron [-] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.088180] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.090845] env[62552]: DEBUG nova.compute.manager [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Stashing vm_state: active {{(pid=62552) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 847.180504] env[62552]: DEBUG nova.compute.manager [req-efab5bf3-ed40-4ce8-aa48-e0754a59a279 req-6d6b5b46-cee8-490f-a49b-5b99cb93ef79 service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Received event network-vif-deleted-810c9b00-6705-4e67-b4f6-f711f07f2037 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 847.180753] env[62552]: INFO nova.compute.manager [req-efab5bf3-ed40-4ce8-aa48-e0754a59a279 req-6d6b5b46-cee8-490f-a49b-5b99cb93ef79 service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Neutron deleted interface 810c9b00-6705-4e67-b4f6-f711f07f2037; detaching it from the instance and deleting it from the info cache [ 847.181946] env[62552]: DEBUG nova.network.neutron [req-efab5bf3-ed40-4ce8-aa48-e0754a59a279 req-6d6b5b46-cee8-490f-a49b-5b99cb93ef79 service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.232940] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b616cf-90e7-443a-b73a-161644922a62 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.243526] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e5704f-51ef-4fef-9fcb-9febf227241a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.278939] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4336c3b4-879c-4ee6-8b0b-72276b371f9d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.294058] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b64b95-59c6-4ac6-b3af-db9f0836fd67 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.297647] env[62552]: DEBUG nova.network.neutron [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Successfully created port: 0679d99d-6c79-408f-82a1-9f80978a1c48 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.309124] env[62552]: DEBUG nova.compute.provider_tree [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 847.318751] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56292} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.319013] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 4083ee43-ecea-4ea5-8923-42b348893824/4083ee43-ecea-4ea5-8923-42b348893824.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.319291] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.319535] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df982b83-b213-4dd5-9dba-165bc4ce1f0e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.328611] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 847.328611] env[62552]: value = "task-1239500" [ 847.328611] env[62552]: _type = "Task" [ 847.328611] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.337016] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239500, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.508415] env[62552]: DEBUG nova.compute.manager [req-ef2ca68c-85bb-4d6c-bce3-9f00ae1752db req-1b779810-a2d2-40f3-9c78-84b2bf3e971f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Received event network-changed-e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 847.508629] env[62552]: DEBUG nova.compute.manager [req-ef2ca68c-85bb-4d6c-bce3-9f00ae1752db req-1b779810-a2d2-40f3-9c78-84b2bf3e971f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Refreshing instance network info cache due to event network-changed-e3439c52-55fc-4ea8-ae30-607c154031cb. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 847.508829] env[62552]: DEBUG oslo_concurrency.lockutils [req-ef2ca68c-85bb-4d6c-bce3-9f00ae1752db req-1b779810-a2d2-40f3-9c78-84b2bf3e971f service nova] Acquiring lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.543238] env[62552]: DEBUG nova.network.neutron [-] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.556021] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Releasing lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.556021] env[62552]: DEBUG nova.compute.manager [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Inject network info {{(pid=62552) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7562}} [ 847.556021] env[62552]: DEBUG nova.compute.manager [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] network_info to inject: |[{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7563}} [ 847.558063] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Reconfiguring VM instance to set the machine id {{(pid=62552) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 847.558352] env[62552]: DEBUG oslo_concurrency.lockutils [req-ef2ca68c-85bb-4d6c-bce3-9f00ae1752db req-1b779810-a2d2-40f3-9c78-84b2bf3e971f service nova] Acquired lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.558534] env[62552]: DEBUG nova.network.neutron [req-ef2ca68c-85bb-4d6c-bce3-9f00ae1752db req-1b779810-a2d2-40f3-9c78-84b2bf3e971f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Refreshing network info cache for port e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.562829] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6141c7f0-0388-4fe6-bba9-617f2e98667b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.579559] env[62552]: DEBUG oslo_vmware.api [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 847.579559] env[62552]: value = "task-1239501" [ 847.579559] env[62552]: _type = "Task" [ 847.579559] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.587291] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239499, 'name': ReconfigVM_Task, 'duration_secs': 0.92343} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.587884] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Reconfigured VM instance instance-00000040 to attach disk [datastore2] d8891025-5bdf-4dc9-a2b0-c86f94582ac6/d8891025-5bdf-4dc9-a2b0-c86f94582ac6.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.589097] env[62552]: INFO nova.compute.manager [-] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Took 1.27 seconds to deallocate network for instance. [ 847.589239] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-478391f5-5a4a-4274-b173-6579f5d7ec76 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.595750] env[62552]: DEBUG oslo_vmware.api [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239501, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.603883] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 847.603883] env[62552]: value = "task-1239502" [ 847.603883] env[62552]: _type = "Task" [ 847.603883] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.612771] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239502, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.616524] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.684690] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3df433ce-d8eb-4cef-ab58-a191272cbd81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.695680] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2222da78-30d0-431c-b839-b0e550ed6a64 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.727035] env[62552]: DEBUG nova.compute.manager [req-efab5bf3-ed40-4ce8-aa48-e0754a59a279 req-6d6b5b46-cee8-490f-a49b-5b99cb93ef79 service nova] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Detach interface failed, port_id=810c9b00-6705-4e67-b4f6-f711f07f2037, reason: Instance c52efdeb-9efc-4ba1-b9af-4dd0a925517f could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 847.787584] env[62552]: DEBUG nova.compute.manager [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 847.815323] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.815599] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.815760] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.815944] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.816118] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.816274] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.816481] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.816641] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.816810] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.816972] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.817176] env[62552]: DEBUG nova.virt.hardware [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.821575] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5159af20-9096-4da3-9809-a71ddb14146b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.834231] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1803379-d4fb-4a81-afd9-83c91fd0eca2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.840849] env[62552]: ERROR nova.scheduler.client.report [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [req-04fdd0cc-d837-43d7-986b-5aa23f5d8e60] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-04fdd0cc-d837-43d7-986b-5aa23f5d8e60"}]} [ 847.847914] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239500, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.138486} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.858951] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.860435] env[62552]: DEBUG nova.scheduler.client.report [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 847.863156] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae717ff-f461-41bb-af1c-cbb40e420d08 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.885513] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 4083ee43-ecea-4ea5-8923-42b348893824/4083ee43-ecea-4ea5-8923-42b348893824.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.886683] env[62552]: DEBUG nova.scheduler.client.report [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 847.886923] env[62552]: DEBUG nova.compute.provider_tree [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 847.888991] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fb37f2c-82ea-4afa-ab81-f066a619536e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.908666] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 847.908666] env[62552]: value = "task-1239503" [ 847.908666] env[62552]: _type = "Task" [ 847.908666] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.912844] env[62552]: DEBUG nova.scheduler.client.report [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 847.917939] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239503, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.935892] env[62552]: DEBUG nova.scheduler.client.report [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 848.049714] env[62552]: INFO nova.compute.manager [-] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Took 1.66 seconds to deallocate network for instance. [ 848.092025] env[62552]: DEBUG oslo_vmware.api [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239501, 'name': ReconfigVM_Task, 'duration_secs': 0.156883} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.092120] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0d36fd59-4e4e-4ba6-9c28-5f655868082c tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Reconfigured VM instance to set the machine id {{(pid=62552) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 848.103498] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.115306] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239502, 'name': Rename_Task, 'duration_secs': 0.416368} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.115573] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.115886] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0cb9e75b-2c90-4a6c-8615-116631beb45b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.121532] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 848.121532] env[62552]: value = "task-1239504" [ 848.121532] env[62552]: _type = "Task" [ 848.121532] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.129873] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239504, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.140573] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.140844] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.141097] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.141653] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.141653] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.144630] env[62552]: INFO nova.compute.manager [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Terminating instance [ 848.347853] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ca69e3-6f08-40a0-8327-aa380656034b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.355480] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e388d7-3a12-4001-a78b-78418745d5d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.386824] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55d3ee4-ac3d-4906-be1c-aba6bbf210f0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.397281] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1512c2-5200-441c-bb57-0ac6bc504009 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.415746] env[62552]: DEBUG nova.compute.provider_tree [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 848.428026] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239503, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.439548] env[62552]: DEBUG nova.network.neutron [req-ef2ca68c-85bb-4d6c-bce3-9f00ae1752db req-1b779810-a2d2-40f3-9c78-84b2bf3e971f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updated VIF entry in instance network info cache for port e3439c52-55fc-4ea8-ae30-607c154031cb. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 848.439902] env[62552]: DEBUG nova.network.neutron [req-ef2ca68c-85bb-4d6c-bce3-9f00ae1752db req-1b779810-a2d2-40f3-9c78-84b2bf3e971f service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updating instance_info_cache with network_info: [{"id": "e3439c52-55fc-4ea8-ae30-607c154031cb", "address": "fa:16:3e:87:99:2b", "network": {"id": "593876e5-aad0-47d9-a797-874e5a569a2b", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-224374093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c4aad184cf448d4be4c03e9e6f3e286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3439c52-55", "ovs_interfaceid": "e3439c52-55fc-4ea8-ae30-607c154031cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.557770] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.632271] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239504, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.652178] env[62552]: DEBUG nova.compute.manager [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 848.652418] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 848.653293] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06b480d-2f06-4ecc-8b68-0b77efd10687 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.660085] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 848.660321] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba7a0dcf-f962-4717-ab71-a851013de61c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.665528] env[62552]: DEBUG oslo_vmware.api [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 848.665528] env[62552]: value = "task-1239505" [ 848.665528] env[62552]: _type = "Task" [ 848.665528] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.673300] env[62552]: DEBUG oslo_vmware.api [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239505, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.927861] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239503, 'name': ReconfigVM_Task, 'duration_secs': 0.856425} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.927861] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 4083ee43-ecea-4ea5-8923-42b348893824/4083ee43-ecea-4ea5-8923-42b348893824.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.927861] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a10e2dc4-b99b-4fd8-8b4a-ad67d6c78927 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.934131] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 848.934131] env[62552]: value = "task-1239506" [ 848.934131] env[62552]: _type = "Task" [ 848.934131] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.943731] env[62552]: DEBUG oslo_concurrency.lockutils [req-ef2ca68c-85bb-4d6c-bce3-9f00ae1752db req-1b779810-a2d2-40f3-9c78-84b2bf3e971f service nova] Releasing lock "refresh_cache-51bbcc4e-8251-4b38-9d36-8aea8fc7705d" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.943998] env[62552]: DEBUG nova.compute.manager [req-ef2ca68c-85bb-4d6c-bce3-9f00ae1752db req-1b779810-a2d2-40f3-9c78-84b2bf3e971f service nova] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Received event network-vif-deleted-8e7f9aff-c687-4cb1-9d2a-fc2b3aaae28f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 848.944358] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239506, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.953282] env[62552]: DEBUG nova.scheduler.client.report [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 91 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 848.956019] env[62552]: DEBUG nova.compute.provider_tree [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 91 to 92 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 848.956019] env[62552]: DEBUG nova.compute.provider_tree [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 849.132890] env[62552]: DEBUG oslo_vmware.api [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239504, 'name': PowerOnVM_Task, 'duration_secs': 0.781718} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.133189] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.133384] env[62552]: INFO nova.compute.manager [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Took 9.40 seconds to spawn the instance on the hypervisor. [ 849.133565] env[62552]: DEBUG nova.compute.manager [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 849.134340] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e572b8-772c-4019-a73f-47f4584507fc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.175715] env[62552]: DEBUG oslo_vmware.api [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239505, 'name': PowerOffVM_Task, 'duration_secs': 0.190862} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.176540] env[62552]: DEBUG nova.network.neutron [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Successfully updated port: 0679d99d-6c79-408f-82a1-9f80978a1c48 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.177737] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 849.177925] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 849.178426] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4045976a-7b74-489b-9cf1-e6a819f575c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.243130] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 849.243468] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 849.243733] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Deleting the datastore file [datastore1] 51bbcc4e-8251-4b38-9d36-8aea8fc7705d {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.244101] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2cd2559-bbc5-499a-937f-ddb9182344cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.251361] env[62552]: DEBUG oslo_vmware.api [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for the task: (returnval){ [ 849.251361] env[62552]: value = "task-1239508" [ 849.251361] env[62552]: _type = "Task" [ 849.251361] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.259675] env[62552]: DEBUG nova.compute.manager [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-vif-plugged-0679d99d-6c79-408f-82a1-9f80978a1c48 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 849.259929] env[62552]: DEBUG oslo_concurrency.lockutils [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.260177] env[62552]: DEBUG oslo_concurrency.lockutils [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.260367] env[62552]: DEBUG oslo_concurrency.lockutils [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.260558] env[62552]: DEBUG nova.compute.manager [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] No waiting events found dispatching network-vif-plugged-0679d99d-6c79-408f-82a1-9f80978a1c48 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 849.260762] env[62552]: WARNING nova.compute.manager [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received unexpected event network-vif-plugged-0679d99d-6c79-408f-82a1-9f80978a1c48 for instance with vm_state building and task_state spawning. [ 849.260942] env[62552]: DEBUG nova.compute.manager [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-changed-0679d99d-6c79-408f-82a1-9f80978a1c48 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 849.261131] env[62552]: DEBUG nova.compute.manager [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Refreshing instance network info cache due to event network-changed-0679d99d-6c79-408f-82a1-9f80978a1c48. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 849.261330] env[62552]: DEBUG oslo_concurrency.lockutils [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] Acquiring lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.261494] env[62552]: DEBUG oslo_concurrency.lockutils [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] Acquired lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.261664] env[62552]: DEBUG nova.network.neutron [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Refreshing network info cache for port 0679d99d-6c79-408f-82a1-9f80978a1c48 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.266586] env[62552]: DEBUG oslo_vmware.api [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239508, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.443873] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239506, 'name': Rename_Task, 'duration_secs': 0.157216} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.444229] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.444501] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8dd67db2-e9d1-424d-874c-41ed4f1f5deb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.449900] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 849.449900] env[62552]: value = "task-1239509" [ 849.449900] env[62552]: _type = "Task" [ 849.449900] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.457724] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.462560] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.711s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.463064] env[62552]: DEBUG nova.compute.manager [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 849.465621] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.222s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.469163] env[62552]: INFO nova.compute.claims [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.656400] env[62552]: INFO nova.compute.manager [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Took 43.46 seconds to build instance. [ 849.679814] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.762179] env[62552]: DEBUG oslo_vmware.api [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Task: {'id': task-1239508, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195884} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.762624] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.763832] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 849.763832] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 849.763832] env[62552]: INFO nova.compute.manager [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 849.763832] env[62552]: DEBUG oslo.service.loopingcall [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.763832] env[62552]: DEBUG nova.compute.manager [-] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 849.764587] env[62552]: DEBUG nova.network.neutron [-] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 849.843050] env[62552]: DEBUG nova.network.neutron [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.961391] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239509, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.977963] env[62552]: DEBUG nova.compute.utils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.980233] env[62552]: DEBUG nova.compute.manager [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 849.980669] env[62552]: DEBUG nova.network.neutron [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 850.086792] env[62552]: DEBUG nova.network.neutron [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.108750] env[62552]: DEBUG nova.policy [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f5312d61ffe4f16ba9e15e0e0d0a2f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81022b826f084769a961659febdacd4d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.162034] env[62552]: DEBUG oslo_concurrency.lockutils [None req-771abb2f-ff8d-48d2-aef5-284440f3cbc2 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.416s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.462409] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239509, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.481217] env[62552]: DEBUG nova.compute.manager [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 850.591071] env[62552]: DEBUG oslo_concurrency.lockutils [req-a0637291-642c-434a-97ad-e5c40fcfa5ed req-c819485a-4f29-4fb9-92c3-cc0f21f41155 service nova] Releasing lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.591681] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.591853] env[62552]: DEBUG nova.network.neutron [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.668640] env[62552]: DEBUG nova.compute.manager [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 850.759151] env[62552]: DEBUG nova.network.neutron [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Successfully created port: c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.934926] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634270b5-8895-4548-8489-2ebfd05be241 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.943013] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035c4446-b830-4d69-a631-220e269cf0ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.977693] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f76f0c2-e4d8-4f50-975b-fbbaad7fcd68 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.990744] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239509, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.992306] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e650b5dd-da8c-48e5-8f6c-10450897b3e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.006371] env[62552]: DEBUG nova.compute.provider_tree [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.178182] env[62552]: DEBUG nova.network.neutron [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.193259] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.319569] env[62552]: DEBUG nova.network.neutron [-] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.484271] env[62552]: DEBUG oslo_vmware.api [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239509, 'name': PowerOnVM_Task, 'duration_secs': 1.920407} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.485718] env[62552]: DEBUG nova.network.neutron [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [{"id": "0679d99d-6c79-408f-82a1-9f80978a1c48", "address": "fa:16:3e:c1:12:aa", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0679d99d-6c", "ovs_interfaceid": "0679d99d-6c79-408f-82a1-9f80978a1c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.487023] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.487121] env[62552]: INFO nova.compute.manager [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Took 9.16 seconds to spawn the instance on the hypervisor. [ 851.487317] env[62552]: DEBUG nova.compute.manager [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 851.488554] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a23923-e39f-4789-b9a5-b4af8223208c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.492293] env[62552]: DEBUG nova.compute.manager [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 851.509149] env[62552]: DEBUG nova.scheduler.client.report [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 851.522786] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.523200] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.523489] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.523768] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.523952] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.524126] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.524344] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.524513] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.524681] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.524854] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.525044] env[62552]: DEBUG nova.virt.hardware [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.526208] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8ab9ab-2438-4ac3-a318-08414799bbe9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.535790] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8faf3fa7-3182-45a2-a4d6-0eb25713388f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.593557] env[62552]: DEBUG nova.compute.manager [req-424412a3-b1a1-487c-a1e6-d3a4353ffa0b req-aa2fff5e-697f-478e-b9d0-c0fd075c80e8 service nova] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Received event network-vif-deleted-e3439c52-55fc-4ea8-ae30-607c154031cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 851.823337] env[62552]: INFO nova.compute.manager [-] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Took 2.06 seconds to deallocate network for instance. [ 851.988994] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.990183] env[62552]: DEBUG nova.compute.manager [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Instance network_info: |[{"id": "0679d99d-6c79-408f-82a1-9f80978a1c48", "address": "fa:16:3e:c1:12:aa", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0679d99d-6c", "ovs_interfaceid": "0679d99d-6c79-408f-82a1-9f80978a1c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 851.990183] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:12:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0a3f99df-d1bc-4a37-a048-263445d4a7b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0679d99d-6c79-408f-82a1-9f80978a1c48', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.998767] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Creating folder: Project (f2f57f72de0c4adc9754458f4e7040b6). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.999069] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bbbfaa27-959f-4482-b0fa-d1becd3711dc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.010142] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Created folder: Project (f2f57f72de0c4adc9754458f4e7040b6) in parent group-v267339. [ 852.010142] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Creating folder: Instances. Parent ref: group-v267426. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.010142] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44026c91-8143-47fd-bbc7-18798cc7786c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.012015] env[62552]: INFO nova.compute.manager [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Took 30.11 seconds to build instance. [ 852.013312] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.013779] env[62552]: DEBUG nova.compute.manager [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 852.017051] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.581s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.017195] env[62552]: DEBUG nova.objects.instance [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lazy-loading 'resources' on Instance uuid 8707ce4b-677e-4f13-86f8-3e327d19380b {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.024023] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Created folder: Instances in parent group-v267426. [ 852.024023] env[62552]: DEBUG oslo.service.loopingcall [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.024537] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.024756] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1381e023-53d0-4cc4-ae1b-62e5c0f843f5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.046162] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.046162] env[62552]: value = "task-1239512" [ 852.046162] env[62552]: _type = "Task" [ 852.046162] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.054565] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239512, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.330759] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.520213] env[62552]: DEBUG oslo_concurrency.lockutils [None req-60b7cd96-c5ce-4769-a3a1-2d03629c7436 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "4083ee43-ecea-4ea5-8923-42b348893824" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.339s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.520213] env[62552]: DEBUG nova.compute.utils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.527318] env[62552]: DEBUG nova.compute.manager [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Not allocating networking since 'none' was specified. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 852.568195] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239512, 'name': CreateVM_Task, 'duration_secs': 0.350829} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.568195] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 852.571085] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.571245] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.571648] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 852.572938] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53d29f2f-0ef4-40a8-8930-cf0d8b98d814 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.577357] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 852.577357] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5274aa25-68db-ebad-6aaa-fd9f5292e8e1" [ 852.577357] env[62552]: _type = "Task" [ 852.577357] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.588408] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5274aa25-68db-ebad-6aaa-fd9f5292e8e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.901294] env[62552]: DEBUG nova.network.neutron [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Successfully updated port: c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.029018] env[62552]: DEBUG nova.compute.manager [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 853.030505] env[62552]: DEBUG nova.compute.manager [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 853.046224] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aabb942-f18d-45e3-ba11-27448f9b443d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.055373] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a370c28b-b4b1-4abb-9a56-a4f70a95df65 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.091032] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607db707-e6a5-481c-b36a-4ebf8740c8a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.102074] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076e3342-b590-4131-ac9f-676f8051af49 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.105852] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5274aa25-68db-ebad-6aaa-fd9f5292e8e1, 'name': SearchDatastore_Task, 'duration_secs': 0.017703} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.106380] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.106624] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.106853] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.107010] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.107200] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.107852] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4b9a09b-c5ec-47ee-a94a-02a84f71ca29 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.118739] env[62552]: DEBUG nova.compute.provider_tree [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.126631] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.126964] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.128054] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5883ef00-e0f6-4ae3-bf86-afff96c31446 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.133655] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 853.133655] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522f4da0-7f4e-9562-7822-dc1b0ded4f49" [ 853.133655] env[62552]: _type = "Task" [ 853.133655] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.141642] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522f4da0-7f4e-9562-7822-dc1b0ded4f49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.404953] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquiring lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.405184] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquired lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.405286] env[62552]: DEBUG nova.network.neutron [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.552338] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.621275] env[62552]: DEBUG nova.scheduler.client.report [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 853.645044] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522f4da0-7f4e-9562-7822-dc1b0ded4f49, 'name': SearchDatastore_Task, 'duration_secs': 0.015834} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.645819] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74a431aa-9e97-412e-835b-26e8807e8236 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.653220] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 853.653220] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525eadf1-3543-5465-9ddb-17b52a3d62b2" [ 853.653220] env[62552]: _type = "Task" [ 853.653220] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.662875] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525eadf1-3543-5465-9ddb-17b52a3d62b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.686062] env[62552]: DEBUG nova.compute.manager [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Received event network-vif-plugged-c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 853.686062] env[62552]: DEBUG oslo_concurrency.lockutils [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] Acquiring lock "1cfeedac-f71e-42e4-a04f-8a0462c85907-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.686503] env[62552]: DEBUG oslo_concurrency.lockutils [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] Lock "1cfeedac-f71e-42e4-a04f-8a0462c85907-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.686503] env[62552]: DEBUG oslo_concurrency.lockutils [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] Lock "1cfeedac-f71e-42e4-a04f-8a0462c85907-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.686503] env[62552]: DEBUG nova.compute.manager [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] No waiting events found dispatching network-vif-plugged-c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 853.686679] env[62552]: WARNING nova.compute.manager [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Received unexpected event network-vif-plugged-c8ca7f73-e9b3-46c6-863c-f304b521dcbc for instance with vm_state building and task_state spawning. [ 853.687237] env[62552]: DEBUG nova.compute.manager [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Received event network-changed-c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 853.687237] env[62552]: DEBUG nova.compute.manager [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Refreshing instance network info cache due to event network-changed-c8ca7f73-e9b3-46c6-863c-f304b521dcbc. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 853.687237] env[62552]: DEBUG oslo_concurrency.lockutils [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] Acquiring lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.942030] env[62552]: DEBUG nova.network.neutron [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.040285] env[62552]: DEBUG nova.compute.manager [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 854.064768] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.065052] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.065397] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.065727] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.065908] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.066076] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.066692] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.066692] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.066692] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.066873] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.066953] env[62552]: DEBUG nova.virt.hardware [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.067841] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244e0a2d-0146-4d64-9b86-7c6d3d5f3030 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.078482] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd03de8-9091-4301-9d24-91e670943507 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.094053] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.100471] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Creating folder: Project (23bf1ae53c3845e9835c50420fad195e). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.100954] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1b1a0b68-5209-4c90-b4a1-b4d815737194 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.111248] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Created folder: Project (23bf1ae53c3845e9835c50420fad195e) in parent group-v267339. [ 854.111456] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Creating folder: Instances. Parent ref: group-v267429. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.111729] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8a0bed1-051a-47e2-9f9c-71518008aaa7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.120111] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Created folder: Instances in parent group-v267429. [ 854.120424] env[62552]: DEBUG oslo.service.loopingcall [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.120556] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.120762] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91ab3e2f-aea2-4bc8-b753-2471ac87f3b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.134611] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.118s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.137697] env[62552]: DEBUG nova.network.neutron [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Updating instance_info_cache with network_info: [{"id": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "address": "fa:16:3e:9f:03:3a", "network": {"id": "d102013b-c1eb-4cd4-82f5-c4fba213a4ca", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1492403044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81022b826f084769a961659febdacd4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8ca7f73-e9", "ovs_interfaceid": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.138657] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.469s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.138887] env[62552]: DEBUG nova.objects.instance [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lazy-loading 'resources' on Instance uuid 36233c83-1170-4a7a-be0b-6a0e8b139de7 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.144750] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.144750] env[62552]: value = "task-1239515" [ 854.144750] env[62552]: _type = "Task" [ 854.144750] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.152612] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239515, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.169791] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525eadf1-3543-5465-9ddb-17b52a3d62b2, 'name': SearchDatastore_Task, 'duration_secs': 0.009875} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.170115] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.170374] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] e31e3077-04e1-4adb-a0cc-44cf84a89eda/e31e3077-04e1-4adb-a0cc-44cf84a89eda.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.170640] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9443b6a0-1dcb-4d83-bd09-3ab3d46f9086 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.177653] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 854.177653] env[62552]: value = "task-1239516" [ 854.177653] env[62552]: _type = "Task" [ 854.177653] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.185481] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239516, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.512079] env[62552]: DEBUG nova.compute.manager [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 854.513118] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9c87c5-ca1a-4a83-8783-582128f65b35 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.641228] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Releasing lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.641581] env[62552]: DEBUG nova.compute.manager [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Instance network_info: |[{"id": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "address": "fa:16:3e:9f:03:3a", "network": {"id": "d102013b-c1eb-4cd4-82f5-c4fba213a4ca", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1492403044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81022b826f084769a961659febdacd4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8ca7f73-e9", "ovs_interfaceid": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 854.644503] env[62552]: DEBUG oslo_concurrency.lockutils [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] Acquired lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.644703] env[62552]: DEBUG nova.network.neutron [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Refreshing network info cache for port c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.645848] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:03:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afd3feb3-ffcc-4499-a2c2-eb6a48aefde9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8ca7f73-e9b3-46c6-863c-f304b521dcbc', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.655045] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Creating folder: Project (81022b826f084769a961659febdacd4d). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.656140] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-138168e8-903f-4e55-b29c-3fb81802769a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.672525] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239515, 'name': CreateVM_Task, 'duration_secs': 0.275652} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.674200] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8190e90b-0dff-49d7-ad2a-eca636df5fe0 tempest-ServerActionsV293TestJSON-514409382 tempest-ServerActionsV293TestJSON-514409382-project-member] Lock "8707ce4b-677e-4f13-86f8-3e327d19380b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.191s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.675268] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.675419] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Created folder: Project (81022b826f084769a961659febdacd4d) in parent group-v267339. [ 854.675493] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Creating folder: Instances. Parent ref: group-v267432. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.675846] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.676126] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.676454] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.677077] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-332bc21f-8a24-4665-96fb-e82821c0470c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.681741] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-457cb7d7-50bd-40f8-b6a9-60c172011fcc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.691980] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 854.691980] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529613dc-76ef-2c19-d4ff-0fab3a0b2dee" [ 854.691980] env[62552]: _type = "Task" [ 854.691980] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.695384] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239516, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.703377] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Created folder: Instances in parent group-v267432. [ 854.703533] env[62552]: DEBUG oslo.service.loopingcall [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.704303] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.704380] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e1fcb60-0b71-4d23-8af0-182bf3d8d666 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.726223] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529613dc-76ef-2c19-d4ff-0fab3a0b2dee, 'name': SearchDatastore_Task, 'duration_secs': 0.01414} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.727140] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.727491] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.727749] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.727914] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.728123] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.728410] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-674e1102-a530-4132-9fd2-cb9db2b76d85 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.731824] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.731824] env[62552]: value = "task-1239519" [ 854.731824] env[62552]: _type = "Task" [ 854.731824] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.741105] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239519, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.742158] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.742345] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.743073] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a9267bc-1765-4705-b5a5-55d8bb260f4c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.748011] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 854.748011] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d113fa-162c-288d-bc63-32d3d2ee0486" [ 854.748011] env[62552]: _type = "Task" [ 854.748011] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.755809] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d113fa-162c-288d-bc63-32d3d2ee0486, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.025403] env[62552]: INFO nova.compute.manager [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] instance snapshotting [ 855.028330] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31f90d5-73dd-4724-a5a3-f7e5860b52ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.054765] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1cde27-80f9-402c-b9eb-00022be43566 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.083614] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e27fd5-c6db-4a01-8af0-3cfda6280e4d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.091494] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26629a3-d4be-45d6-b243-ce1f7a10b99a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.122205] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6022fb-8a65-4f2e-a8d2-4db9c9fdfa80 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.130137] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364b1ef1-daf3-4173-8208-54bf54dbdd7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.143369] env[62552]: DEBUG nova.compute.provider_tree [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.194445] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239516, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518964} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.194726] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] e31e3077-04e1-4adb-a0cc-44cf84a89eda/e31e3077-04e1-4adb-a0cc-44cf84a89eda.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 855.194945] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.195216] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c7609959-8775-4485-8176-7f446dcbdeda {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.201313] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 855.201313] env[62552]: value = "task-1239520" [ 855.201313] env[62552]: _type = "Task" [ 855.201313] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.210271] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239520, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.242910] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239519, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.258716] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d113fa-162c-288d-bc63-32d3d2ee0486, 'name': SearchDatastore_Task, 'duration_secs': 0.062221} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.259673] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd49053c-a43f-42f6-97c5-b1ba113fd48e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.265254] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 855.265254] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cf1aa2-42d8-d99d-01ee-dd6ab031573b" [ 855.265254] env[62552]: _type = "Task" [ 855.265254] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.273961] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cf1aa2-42d8-d99d-01ee-dd6ab031573b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.516058] env[62552]: DEBUG nova.network.neutron [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Updated VIF entry in instance network info cache for port c8ca7f73-e9b3-46c6-863c-f304b521dcbc. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.516459] env[62552]: DEBUG nova.network.neutron [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Updating instance_info_cache with network_info: [{"id": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "address": "fa:16:3e:9f:03:3a", "network": {"id": "d102013b-c1eb-4cd4-82f5-c4fba213a4ca", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1492403044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81022b826f084769a961659febdacd4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8ca7f73-e9", "ovs_interfaceid": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.564989] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 855.565328] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d63bf01f-e54d-465c-bad4-543f04ecd70b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.572934] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 855.572934] env[62552]: value = "task-1239521" [ 855.572934] env[62552]: _type = "Task" [ 855.572934] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.584699] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239521, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.646490] env[62552]: DEBUG nova.scheduler.client.report [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 855.711101] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239520, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062545} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.711511] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.712376] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edbdcdf-8fd0-4b61-a69d-c95b5ec9a1e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.735178] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] e31e3077-04e1-4adb-a0cc-44cf84a89eda/e31e3077-04e1-4adb-a0cc-44cf84a89eda.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.736107] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdc6869f-f7dc-4f89-8bd4-fcfe83cf60d1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.772698] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239519, 'name': CreateVM_Task, 'duration_secs': 0.849126} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.773034] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 855.773034] env[62552]: value = "task-1239522" [ 855.773034] env[62552]: _type = "Task" [ 855.773034] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.773599] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.774439] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.774642] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.775103] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.781304] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7faf15cb-0ea2-440b-9241-fa314ba76ce9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.782800] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cf1aa2-42d8-d99d-01ee-dd6ab031573b, 'name': SearchDatastore_Task, 'duration_secs': 0.015317} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.783363] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.784151] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] fb881c34-dccc-4703-af8e-c75caafd9b08/fb881c34-dccc-4703-af8e-c75caafd9b08.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.784151] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c26ffb7c-647d-4187-90a7-771f1681f7f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.790259] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239522, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.790564] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 855.790564] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527bd742-8614-4b80-6569-9256a17619fa" [ 855.790564] env[62552]: _type = "Task" [ 855.790564] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.794996] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 855.794996] env[62552]: value = "task-1239523" [ 855.794996] env[62552]: _type = "Task" [ 855.794996] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.803579] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527bd742-8614-4b80-6569-9256a17619fa, 'name': SearchDatastore_Task, 'duration_secs': 0.010936} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.804238] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.804500] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.804744] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.804897] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.805091] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.805349] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b06272b-c9ae-4bca-af09-c92897b8400e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.810195] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239523, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.815237] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.815333] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.816285] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96b1a723-a497-4bdc-a9c9-7119b0fe056e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.821092] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 855.821092] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f12e14-a725-e897-0b18-591026c4b921" [ 855.821092] env[62552]: _type = "Task" [ 855.821092] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.828649] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f12e14-a725-e897-0b18-591026c4b921, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.019668] env[62552]: DEBUG oslo_concurrency.lockutils [req-8d4d5a05-0ac7-4b79-94bc-d82392a1304e req-554c3d07-ab78-48c8-880a-7a5bf7edac92 service nova] Releasing lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.087147] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239521, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.151783] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.013s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.155636] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.253s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.157423] env[62552]: INFO nova.compute.claims [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.183664] env[62552]: INFO nova.scheduler.client.report [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Deleted allocations for instance 36233c83-1170-4a7a-be0b-6a0e8b139de7 [ 856.284029] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239522, 'name': ReconfigVM_Task, 'duration_secs': 0.306436} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.284029] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Reconfigured VM instance instance-00000042 to attach disk [datastore2] e31e3077-04e1-4adb-a0cc-44cf84a89eda/e31e3077-04e1-4adb-a0cc-44cf84a89eda.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.284258] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b9c7a32-09f6-4a7f-80d5-009bfe695c36 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.290539] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 856.290539] env[62552]: value = "task-1239524" [ 856.290539] env[62552]: _type = "Task" [ 856.290539] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.298264] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239524, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.305354] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239523, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458676} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.305584] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] fb881c34-dccc-4703-af8e-c75caafd9b08/fb881c34-dccc-4703-af8e-c75caafd9b08.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.305797] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.306048] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2c4159c-95a7-4bb1-86a4-032bf16a3c94 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.312185] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 856.312185] env[62552]: value = "task-1239525" [ 856.312185] env[62552]: _type = "Task" [ 856.312185] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.319675] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239525, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.329674] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f12e14-a725-e897-0b18-591026c4b921, 'name': SearchDatastore_Task, 'duration_secs': 0.00944} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.330453] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01f27d42-a521-40b3-8846-1b00771c6a18 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.337023] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 856.337023] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a50881-3675-9dc5-43fa-2230f216bfdf" [ 856.337023] env[62552]: _type = "Task" [ 856.337023] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.342221] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a50881-3675-9dc5-43fa-2230f216bfdf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.584137] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239521, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.697146] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1dd5cde-3c98-4100-948f-c997f8fe67e2 tempest-ServersTestMultiNic-555568841 tempest-ServersTestMultiNic-555568841-project-member] Lock "36233c83-1170-4a7a-be0b-6a0e8b139de7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.560s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.801801] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239524, 'name': Rename_Task, 'duration_secs': 0.141781} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.802603] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.802875] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c99be027-8424-4f71-a75e-a487482ba049 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.808511] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 856.808511] env[62552]: value = "task-1239526" [ 856.808511] env[62552]: _type = "Task" [ 856.808511] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.819085] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239526, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.824073] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239525, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07381} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.824358] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.825157] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2edff3-d5ff-407d-87a4-7ee3d89ab6a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.849599] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] fb881c34-dccc-4703-af8e-c75caafd9b08/fb881c34-dccc-4703-af8e-c75caafd9b08.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.852991] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29cdcad0-27ca-4084-bc70-a2fd59a4ebe7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.877191] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a50881-3675-9dc5-43fa-2230f216bfdf, 'name': SearchDatastore_Task, 'duration_secs': 0.008341} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.878652] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.878940] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 1cfeedac-f71e-42e4-a04f-8a0462c85907/1cfeedac-f71e-42e4-a04f-8a0462c85907.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.879343] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 856.879343] env[62552]: value = "task-1239527" [ 856.879343] env[62552]: _type = "Task" [ 856.879343] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.879552] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a9f674b-4343-40a4-836b-3dc8521c1440 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.890402] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239527, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.891784] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 856.891784] env[62552]: value = "task-1239528" [ 856.891784] env[62552]: _type = "Task" [ 856.891784] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.900463] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239528, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.085592] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239521, 'name': CreateSnapshot_Task, 'duration_secs': 1.154745} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.085873] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 857.086793] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3aaadda-42a0-49fd-ac17-9352900b236a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.319763] env[62552]: DEBUG oslo_vmware.api [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239526, 'name': PowerOnVM_Task, 'duration_secs': 0.504723} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.321339] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.321339] env[62552]: INFO nova.compute.manager [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Took 9.53 seconds to spawn the instance on the hypervisor. [ 857.321339] env[62552]: DEBUG nova.compute.manager [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 857.321533] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7f371e-3f71-4e02-857d-e5a284587efb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.396902] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239527, 'name': ReconfigVM_Task, 'duration_secs': 0.34083} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.404560] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Reconfigured VM instance instance-00000044 to attach disk [datastore2] fb881c34-dccc-4703-af8e-c75caafd9b08/fb881c34-dccc-4703-af8e-c75caafd9b08.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.406696] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70dacee1-9037-4783-9561-0c7a4b44f078 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.414053] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239528, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.416618] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 857.416618] env[62552]: value = "task-1239529" [ 857.416618] env[62552]: _type = "Task" [ 857.416618] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.425267] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239529, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.606388] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 857.609162] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5d4d30a0-e8a3-4de8-abbb-56e68e6679f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.617886] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 857.617886] env[62552]: value = "task-1239530" [ 857.617886] env[62552]: _type = "Task" [ 857.617886] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.630475] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239530, 'name': CloneVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.695066] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e33ebb-9de1-4d4d-8987-a6b1edab1604 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.703853] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9181e0d5-ab64-4691-8e2f-62e4e861a11c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.740052] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf23406-9448-46b7-a95c-275c1f6aafcf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.747942] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1f71f1-fe1c-4c1c-8dd8-bb39032c82ea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.763202] env[62552]: DEBUG nova.compute.provider_tree [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 857.842820] env[62552]: INFO nova.compute.manager [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Took 33.85 seconds to build instance. [ 857.908939] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239528, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649382} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.909271] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 1cfeedac-f71e-42e4-a04f-8a0462c85907/1cfeedac-f71e-42e4-a04f-8a0462c85907.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 857.909515] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.909837] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-80ab3b43-fd6b-491f-8899-9ccb5e9dbdb1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.918363] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 857.918363] env[62552]: value = "task-1239531" [ 857.918363] env[62552]: _type = "Task" [ 857.918363] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.929988] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239529, 'name': Rename_Task, 'duration_secs': 0.355805} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.932983] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.933270] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239531, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.933490] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eea23410-c257-4787-b6ba-a322b9efb1db {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.940124] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 857.940124] env[62552]: value = "task-1239532" [ 857.940124] env[62552]: _type = "Task" [ 857.940124] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.948636] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239532, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.130154] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239530, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.305192] env[62552]: DEBUG nova.scheduler.client.report [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 92 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 858.305566] env[62552]: DEBUG nova.compute.provider_tree [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 92 to 93 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 858.306185] env[62552]: DEBUG nova.compute.provider_tree [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 858.345232] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e035bb9-036d-43cd-b114-67e6268eefe6 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.767s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.429936] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239531, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.129894} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.430245] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.431117] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58ea3c4-3709-436f-89e2-20e2951a21b6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.457118] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 1cfeedac-f71e-42e4-a04f-8a0462c85907/1cfeedac-f71e-42e4-a04f-8a0462c85907.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.460872] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba49669c-a413-424b-8386-0791d3c6ead9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.499256] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239532, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.500959] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 858.500959] env[62552]: value = "task-1239533" [ 858.500959] env[62552]: _type = "Task" [ 858.500959] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.508912] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239533, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.630667] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239530, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.812275] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.656s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.813136] env[62552]: DEBUG nova.compute.manager [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 858.818056] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.289s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.818056] env[62552]: DEBUG nova.objects.instance [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lazy-loading 'resources' on Instance uuid fef94e30-f946-4d7f-a108-f32cad505bc4 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.843020] env[62552]: DEBUG nova.compute.manager [req-ba0bfb98-3aec-4870-9891-a32d5f088d2e req-5fc8ba47-e3a2-47f7-b192-961831b587f2 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-changed-0679d99d-6c79-408f-82a1-9f80978a1c48 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 858.843020] env[62552]: DEBUG nova.compute.manager [req-ba0bfb98-3aec-4870-9891-a32d5f088d2e req-5fc8ba47-e3a2-47f7-b192-961831b587f2 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Refreshing instance network info cache due to event network-changed-0679d99d-6c79-408f-82a1-9f80978a1c48. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 858.843020] env[62552]: DEBUG oslo_concurrency.lockutils [req-ba0bfb98-3aec-4870-9891-a32d5f088d2e req-5fc8ba47-e3a2-47f7-b192-961831b587f2 service nova] Acquiring lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.843020] env[62552]: DEBUG oslo_concurrency.lockutils [req-ba0bfb98-3aec-4870-9891-a32d5f088d2e req-5fc8ba47-e3a2-47f7-b192-961831b587f2 service nova] Acquired lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.843020] env[62552]: DEBUG nova.network.neutron [req-ba0bfb98-3aec-4870-9891-a32d5f088d2e req-5fc8ba47-e3a2-47f7-b192-961831b587f2 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Refreshing network info cache for port 0679d99d-6c79-408f-82a1-9f80978a1c48 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.852076] env[62552]: DEBUG nova.compute.manager [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 858.960577] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239532, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.013843] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239533, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.132964] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239530, 'name': CloneVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.323550] env[62552]: DEBUG nova.compute.utils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.331690] env[62552]: DEBUG nova.compute.manager [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Not allocating networking since 'none' was specified. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 859.381912] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.459376] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239532, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.516645] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239533, 'name': ReconfigVM_Task, 'duration_secs': 0.535173} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.518740] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 1cfeedac-f71e-42e4-a04f-8a0462c85907/1cfeedac-f71e-42e4-a04f-8a0462c85907.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.520071] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-acca1d25-d275-4ee1-9b7c-45d299b5c867 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.528200] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 859.528200] env[62552]: value = "task-1239534" [ 859.528200] env[62552]: _type = "Task" [ 859.528200] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.540561] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239534, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.634930] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239530, 'name': CloneVM_Task, 'duration_secs': 1.582738} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.634930] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Created linked-clone VM from snapshot [ 859.634930] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1134f5-a25f-403a-a35b-fbe2b394f53f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.641755] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Uploading image 967e9aaa-56d6-402c-abae-6f8af6ce99e2 {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 859.671409] env[62552]: DEBUG oslo_vmware.rw_handles [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 859.671409] env[62552]: value = "vm-267436" [ 859.671409] env[62552]: _type = "VirtualMachine" [ 859.671409] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 859.672047] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c8830e93-dab9-4a5b-84c7-dc932c6d090e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.684727] env[62552]: DEBUG oslo_vmware.rw_handles [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lease: (returnval){ [ 859.684727] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52850828-c1dc-4d8b-f982-67b99b1431a7" [ 859.684727] env[62552]: _type = "HttpNfcLease" [ 859.684727] env[62552]: } obtained for exporting VM: (result){ [ 859.684727] env[62552]: value = "vm-267436" [ 859.684727] env[62552]: _type = "VirtualMachine" [ 859.684727] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 859.684727] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the lease: (returnval){ [ 859.684727] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52850828-c1dc-4d8b-f982-67b99b1431a7" [ 859.684727] env[62552]: _type = "HttpNfcLease" [ 859.684727] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 859.688156] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 859.688156] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52850828-c1dc-4d8b-f982-67b99b1431a7" [ 859.688156] env[62552]: _type = "HttpNfcLease" [ 859.688156] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 859.762335] env[62552]: DEBUG nova.network.neutron [req-ba0bfb98-3aec-4870-9891-a32d5f088d2e req-5fc8ba47-e3a2-47f7-b192-961831b587f2 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updated VIF entry in instance network info cache for port 0679d99d-6c79-408f-82a1-9f80978a1c48. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.762759] env[62552]: DEBUG nova.network.neutron [req-ba0bfb98-3aec-4870-9891-a32d5f088d2e req-5fc8ba47-e3a2-47f7-b192-961831b587f2 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [{"id": "0679d99d-6c79-408f-82a1-9f80978a1c48", "address": "fa:16:3e:c1:12:aa", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0679d99d-6c", "ovs_interfaceid": "0679d99d-6c79-408f-82a1-9f80978a1c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.826766] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2607b7c4-4cf0-453e-b0de-6217f4ce48c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.831702] env[62552]: DEBUG nova.compute.manager [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 859.840285] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9845a9b8-7fb1-47b7-bc16-554737669850 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.873775] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3658e75d-dd86-4599-b059-59970aca2ebb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.881608] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de3a938-6437-4633-8ef6-7b7326bdaa5a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.896134] env[62552]: DEBUG nova.compute.provider_tree [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.957439] env[62552]: DEBUG oslo_vmware.api [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239532, 'name': PowerOnVM_Task, 'duration_secs': 1.883496} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.957713] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 859.957932] env[62552]: INFO nova.compute.manager [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Took 5.92 seconds to spawn the instance on the hypervisor. [ 859.958130] env[62552]: DEBUG nova.compute.manager [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 859.958951] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41dce8e-c847-47fd-ac1e-a4e0d152c82a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.040475] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239534, 'name': Rename_Task, 'duration_secs': 0.150652} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.040806] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.041076] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b04c6a1b-8665-4f57-854e-3b70811601b5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.047495] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 860.047495] env[62552]: value = "task-1239536" [ 860.047495] env[62552]: _type = "Task" [ 860.047495] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.054801] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.189512] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 860.189512] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52850828-c1dc-4d8b-f982-67b99b1431a7" [ 860.189512] env[62552]: _type = "HttpNfcLease" [ 860.189512] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 860.189911] env[62552]: DEBUG oslo_vmware.rw_handles [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 860.189911] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52850828-c1dc-4d8b-f982-67b99b1431a7" [ 860.189911] env[62552]: _type = "HttpNfcLease" [ 860.189911] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 860.190753] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b77bd78-f392-4f11-9ec5-039306d0eea8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.197794] env[62552]: DEBUG oslo_vmware.rw_handles [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b591c7-af48-100d-415e-a3a0906e6d1c/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 860.198037] env[62552]: DEBUG oslo_vmware.rw_handles [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b591c7-af48-100d-415e-a3a0906e6d1c/disk-0.vmdk for reading. {{(pid=62552) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 860.267361] env[62552]: DEBUG oslo_concurrency.lockutils [req-ba0bfb98-3aec-4870-9891-a32d5f088d2e req-5fc8ba47-e3a2-47f7-b192-961831b587f2 service nova] Releasing lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.287226] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c9e921ac-ef8a-43cf-b6aa-c834740b218c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.399866] env[62552]: DEBUG nova.scheduler.client.report [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 860.477166] env[62552]: INFO nova.compute.manager [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Took 32.25 seconds to build instance. [ 860.559421] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239536, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.843245] env[62552]: DEBUG nova.compute.manager [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 860.877013] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.877413] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.877735] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.878022] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.878239] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.878436] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.878825] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.879142] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.879398] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.879725] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.879960] env[62552]: DEBUG nova.virt.hardware [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.881015] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1663bddb-4940-4b3c-acd1-ab30034cccae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.890137] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4131963-2ad3-4b2e-8573-cd829b719193 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.905600] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.089s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.909609] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.916497] env[62552]: DEBUG oslo.service.loopingcall [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.917309] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.587s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.917548] env[62552]: DEBUG nova.objects.instance [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lazy-loading 'resources' on Instance uuid ce5d0165-65f1-4505-9c46-1129c56a8913 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.919090] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 860.919447] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ee63fb0-1941-4bae-b0a2-bb0aab9aab03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.942523] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.942523] env[62552]: value = "task-1239537" [ 860.942523] env[62552]: _type = "Task" [ 860.942523] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.946572] env[62552]: INFO nova.scheduler.client.report [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Deleted allocations for instance fef94e30-f946-4d7f-a108-f32cad505bc4 [ 860.950956] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239537, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.979339] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0758a570-9a61-43f0-ae47-a91588354409 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fb881c34-dccc-4703-af8e-c75caafd9b08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.920s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.062555] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239536, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.382239] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bc5c83-eaf8-4744-8b4b-148cf84d8a53 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.392500] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db20ef70-7640-43fa-99c8-58c02bb98b00 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.425457] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04f06d0-6e6b-4e4e-900d-5e78ae636b82 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.435048] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781bb2d6-8dae-47d3-911d-4b109e9edc23 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.449620] env[62552]: DEBUG nova.compute.provider_tree [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.460821] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239537, 'name': CreateVM_Task, 'duration_secs': 0.302373} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.461404] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f7c9238-33be-4a1d-9139-38e015479256 tempest-InstanceActionsNegativeTestJSON-243170549 tempest-InstanceActionsNegativeTestJSON-243170549-project-member] Lock "fef94e30-f946-4d7f-a108-f32cad505bc4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.796s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.462286] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 861.462938] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.463318] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.463655] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 861.463892] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be7bdf08-ce2d-4d25-b7b1-cb31ade2b3a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.469859] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 861.469859] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5290d97b-82c8-b576-3dec-82a410bf9bd5" [ 861.469859] env[62552]: _type = "Task" [ 861.469859] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.487145] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5290d97b-82c8-b576-3dec-82a410bf9bd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.487145] env[62552]: DEBUG nova.compute.manager [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 861.559777] env[62552]: DEBUG oslo_vmware.api [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239536, 'name': PowerOnVM_Task, 'duration_secs': 1.209042} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.559987] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.560254] env[62552]: INFO nova.compute.manager [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Took 10.07 seconds to spawn the instance on the hypervisor. [ 861.560475] env[62552]: DEBUG nova.compute.manager [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 861.561511] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f9c091-307e-42cc-978b-7e7112bf9afa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.955770] env[62552]: DEBUG nova.scheduler.client.report [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 861.981687] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5290d97b-82c8-b576-3dec-82a410bf9bd5, 'name': SearchDatastore_Task, 'duration_secs': 0.011599} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.981997] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.982265] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.982504] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.982652] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.982837] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.983121] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f82e2b1d-ce8d-412b-808f-5ee4a0c5b1bd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.992729] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.992921] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 861.996105] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73d05a31-5808-4f94-ab29-84c35e58560d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.004480] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 862.004480] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52305026-92f4-32e6-b4bc-94a599cdfea0" [ 862.004480] env[62552]: _type = "Task" [ 862.004480] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.012662] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52305026-92f4-32e6-b4bc-94a599cdfea0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.014957] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.081975] env[62552]: INFO nova.compute.manager [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Took 36.39 seconds to build instance. [ 862.460585] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.543s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.467656] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.397s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.467656] env[62552]: INFO nova.compute.claims [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.491029] env[62552]: INFO nova.scheduler.client.report [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Deleted allocations for instance ce5d0165-65f1-4505-9c46-1129c56a8913 [ 862.515901] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52305026-92f4-32e6-b4bc-94a599cdfea0, 'name': SearchDatastore_Task, 'duration_secs': 0.013585} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.516901] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-865ff23d-d6e2-4330-b8b1-620a43c0bdd2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.522734] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 862.522734] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526d47ab-75f7-2b08-bf68-adb04e9674ab" [ 862.522734] env[62552]: _type = "Task" [ 862.522734] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.532107] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526d47ab-75f7-2b08-bf68-adb04e9674ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.584101] env[62552]: DEBUG oslo_concurrency.lockutils [None req-54a8ecb2-1ddb-4b74-9ad6-618d19fe64ae tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "1cfeedac-f71e-42e4-a04f-8a0462c85907" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.623s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.005808] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6ac5330c-ba3d-4ad6-9ef0-37c31aa3c60c tempest-ServersTestBootFromVolume-934679892 tempest-ServersTestBootFromVolume-934679892-project-member] Lock "ce5d0165-65f1-4505-9c46-1129c56a8913" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.770s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.034161] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526d47ab-75f7-2b08-bf68-adb04e9674ab, 'name': SearchDatastore_Task, 'duration_secs': 0.014128} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.034477] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.034696] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] fea83dde-3181-49dd-a000-5e5cffc8de95/fea83dde-3181-49dd-a000-5e5cffc8de95.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 863.034963] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-176ece37-bb53-4710-96ce-5d681a106565 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.046179] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 863.046179] env[62552]: value = "task-1239538" [ 863.046179] env[62552]: _type = "Task" [ 863.046179] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.056717] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.086643] env[62552]: DEBUG nova.compute.manager [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 863.558508] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239538, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.608314] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.971697] env[62552]: DEBUG nova.compute.manager [req-8c07ebfa-7b72-45ed-86c5-bffce118f4de req-5b01da3f-308e-4e32-b4fe-469ab7bb42dc service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Received event network-changed-c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 863.971697] env[62552]: DEBUG nova.compute.manager [req-8c07ebfa-7b72-45ed-86c5-bffce118f4de req-5b01da3f-308e-4e32-b4fe-469ab7bb42dc service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Refreshing instance network info cache due to event network-changed-c8ca7f73-e9b3-46c6-863c-f304b521dcbc. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 863.973270] env[62552]: DEBUG oslo_concurrency.lockutils [req-8c07ebfa-7b72-45ed-86c5-bffce118f4de req-5b01da3f-308e-4e32-b4fe-469ab7bb42dc service nova] Acquiring lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.973462] env[62552]: DEBUG oslo_concurrency.lockutils [req-8c07ebfa-7b72-45ed-86c5-bffce118f4de req-5b01da3f-308e-4e32-b4fe-469ab7bb42dc service nova] Acquired lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.973734] env[62552]: DEBUG nova.network.neutron [req-8c07ebfa-7b72-45ed-86c5-bffce118f4de req-5b01da3f-308e-4e32-b4fe-469ab7bb42dc service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Refreshing network info cache for port c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.976757] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b239abd-a5f9-4fb8-861c-bfe8916f097c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.990981] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47da1531-9c06-46dd-95aa-c2d386efd9cb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.026475] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04b1ca3-f46e-4c7d-9a31-80152eeebd80 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.035217] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dffea1c-5ba7-451a-9342-ac1294208a4a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.053174] env[62552]: DEBUG nova.compute.provider_tree [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.061064] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239538, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.739445} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.061974] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] fea83dde-3181-49dd-a000-5e5cffc8de95/fea83dde-3181-49dd-a000-5e5cffc8de95.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 864.062253] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.062472] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad02c736-957a-4422-8659-caa6b924b863 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.070454] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 864.070454] env[62552]: value = "task-1239539" [ 864.070454] env[62552]: _type = "Task" [ 864.070454] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.081339] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239539, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.558736] env[62552]: DEBUG nova.scheduler.client.report [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 864.586965] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239539, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125624} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.587327] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.588155] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff230fb6-85fa-4ce5-be20-f24dc7dee5f7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.611684] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] fea83dde-3181-49dd-a000-5e5cffc8de95/fea83dde-3181-49dd-a000-5e5cffc8de95.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.612864] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce4a8828-af92-4050-a53e-79f8d0b2f9c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.634663] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 864.634663] env[62552]: value = "task-1239540" [ 864.634663] env[62552]: _type = "Task" [ 864.634663] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.644315] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239540, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.070747] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.607s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.071901] env[62552]: DEBUG nova.compute.manager [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 865.076539] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.107s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.076766] env[62552]: DEBUG nova.objects.instance [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'resources' on Instance uuid cdee22a9-4327-47af-ab14-8403a06b2802 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.085535] env[62552]: DEBUG nova.network.neutron [req-8c07ebfa-7b72-45ed-86c5-bffce118f4de req-5b01da3f-308e-4e32-b4fe-469ab7bb42dc service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Updated VIF entry in instance network info cache for port c8ca7f73-e9b3-46c6-863c-f304b521dcbc. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.085981] env[62552]: DEBUG nova.network.neutron [req-8c07ebfa-7b72-45ed-86c5-bffce118f4de req-5b01da3f-308e-4e32-b4fe-469ab7bb42dc service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Updating instance_info_cache with network_info: [{"id": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "address": "fa:16:3e:9f:03:3a", "network": {"id": "d102013b-c1eb-4cd4-82f5-c4fba213a4ca", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1492403044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81022b826f084769a961659febdacd4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8ca7f73-e9", "ovs_interfaceid": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.146416] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239540, 'name': ReconfigVM_Task, 'duration_secs': 0.286367} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.146715] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Reconfigured VM instance instance-00000045 to attach disk [datastore1] fea83dde-3181-49dd-a000-5e5cffc8de95/fea83dde-3181-49dd-a000-5e5cffc8de95.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.147392] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-caa58167-1356-4edf-b4cf-f0973ecd82db {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.154220] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 865.154220] env[62552]: value = "task-1239541" [ 865.154220] env[62552]: _type = "Task" [ 865.154220] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.163623] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239541, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.583992] env[62552]: DEBUG nova.objects.instance [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'numa_topology' on Instance uuid cdee22a9-4327-47af-ab14-8403a06b2802 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.590974] env[62552]: DEBUG nova.compute.utils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 865.598360] env[62552]: DEBUG oslo_concurrency.lockutils [req-8c07ebfa-7b72-45ed-86c5-bffce118f4de req-5b01da3f-308e-4e32-b4fe-469ab7bb42dc service nova] Releasing lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.599047] env[62552]: DEBUG nova.compute.manager [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 865.599353] env[62552]: DEBUG nova.network.neutron [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 865.667693] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239541, 'name': Rename_Task, 'duration_secs': 0.141459} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.667693] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.667693] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0f75401-eb95-470d-869f-0968ab3d5e6f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.675019] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 865.675019] env[62552]: value = "task-1239542" [ 865.675019] env[62552]: _type = "Task" [ 865.675019] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.680325] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239542, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.685122] env[62552]: DEBUG nova.policy [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '75e6d5f39c8d46b1bfb89654ff27d821', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51c2a67f74dc42d2b915e0429199f89b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 865.913442] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "80478878-ff82-4ed6-a851-8eb2bec01e22" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.913442] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.100850] env[62552]: DEBUG nova.objects.base [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 866.111093] env[62552]: DEBUG nova.compute.manager [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 866.192018] env[62552]: DEBUG oslo_vmware.api [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239542, 'name': PowerOnVM_Task, 'duration_secs': 0.425169} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.192018] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.192018] env[62552]: INFO nova.compute.manager [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Took 5.35 seconds to spawn the instance on the hypervisor. [ 866.192018] env[62552]: DEBUG nova.compute.manager [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 866.192018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec080fde-364f-45a9-85a9-57b98f21a429 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.336478] env[62552]: DEBUG nova.network.neutron [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Successfully created port: e421f4d6-d892-4c96-8c11-32f1b7553780 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.416971] env[62552]: DEBUG nova.compute.utils [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.645957] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789e3e15-7b77-465e-afb1-0192808ee550 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.653534] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265e3348-4d3f-45d5-b481-4cbd42382e10 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.687108] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa82fff-7e88-408f-a34f-a9be7ce8a751 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.702666] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967b78fe-cd06-4ee2-895d-ca2e452b62c4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.730834] env[62552]: DEBUG nova.compute.provider_tree [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.735867] env[62552]: INFO nova.compute.manager [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Took 36.86 seconds to build instance. [ 866.920461] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.121609] env[62552]: DEBUG nova.compute.manager [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 867.159687] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.160073] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.160461] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.163924] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.163924] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.163924] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.163924] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.163924] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.163924] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.163924] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.163924] env[62552]: DEBUG nova.virt.hardware [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.163924] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2472dfee-6b2f-4eec-bb72-72e8befebe05 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.174162] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d940ca-63ce-44ab-9292-83aadc7130ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.234330] env[62552]: DEBUG nova.scheduler.client.report [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 867.239106] env[62552]: DEBUG oslo_concurrency.lockutils [None req-af8ac57f-d5f8-45fa-bb6b-7118ced382ae tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fea83dde-3181-49dd-a000-5e5cffc8de95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.540s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.718322] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 867.718322] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 867.741936] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.665s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.746143] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.545s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.746438] env[62552]: DEBUG nova.objects.instance [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'resources' on Instance uuid 3b333977-0b4b-4ccf-bed5-231215abae3c {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.001565] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "80478878-ff82-4ed6-a851-8eb2bec01e22" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.001881] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.002169] env[62552]: INFO nova.compute.manager [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Attaching volume f925d139-4314-4651-be8c-f07a21f94eea to /dev/sdb [ 868.052444] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c03a4f-9726-47ef-8f23-840709be135c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.061536] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b331ea-df1a-40cb-aefb-2aa390983db0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.075650] env[62552]: DEBUG nova.virt.block_device [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Updating existing volume attachment record: 71b3d7c8-7aba-4567-80c8-1d2b8265e7e8 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 868.228594] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 868.228909] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 868.228909] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Rebuilding the list of instances to heal {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 868.263420] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a63acc25-08fb-4173-ba08-10a2d176899c tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 55.348s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.266631] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 27.187s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.267259] env[62552]: INFO nova.compute.manager [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Unshelving [ 868.710724] env[62552]: DEBUG nova.compute.manager [req-5379a909-f883-43c3-8237-25588f62dbe8 req-f84c8724-db11-4f40-be1a-4083857a0ca8 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Received event network-changed-c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 868.710999] env[62552]: DEBUG nova.compute.manager [req-5379a909-f883-43c3-8237-25588f62dbe8 req-f84c8724-db11-4f40-be1a-4083857a0ca8 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Refreshing instance network info cache due to event network-changed-c8ca7f73-e9b3-46c6-863c-f304b521dcbc. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 868.712470] env[62552]: DEBUG oslo_concurrency.lockutils [req-5379a909-f883-43c3-8237-25588f62dbe8 req-f84c8724-db11-4f40-be1a-4083857a0ca8 service nova] Acquiring lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.713118] env[62552]: DEBUG oslo_concurrency.lockutils [req-5379a909-f883-43c3-8237-25588f62dbe8 req-f84c8724-db11-4f40-be1a-4083857a0ca8 service nova] Acquired lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.713118] env[62552]: DEBUG nova.network.neutron [req-5379a909-f883-43c3-8237-25588f62dbe8 req-f84c8724-db11-4f40-be1a-4083857a0ca8 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Refreshing network info cache for port c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.738977] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Skipping network cache update for instance because it is Building. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 868.760837] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9db4735-7fbc-4a8a-8f07-61b2a9a63c1f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.773293] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6632736c-90b1-4104-9152-33925f2194e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.825899] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.825899] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquired lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.825899] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Forcefully refreshing network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 868.825899] env[62552]: DEBUG nova.objects.instance [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lazy-loading 'info_cache' on Instance uuid 0418260a-aa27-4955-ab15-b180ec04f0b3 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.828203] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b87bc3-71e7-4705-adc1-22903310e314 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.840802] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33aec3d0-9158-465c-849e-aa658d7f8d92 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.861198] env[62552]: DEBUG nova.compute.provider_tree [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.998557] env[62552]: DEBUG nova.network.neutron [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Successfully updated port: e421f4d6-d892-4c96-8c11-32f1b7553780 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.286209] env[62552]: DEBUG nova.compute.utils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.368029] env[62552]: DEBUG nova.scheduler.client.report [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 869.505244] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquiring lock "refresh_cache-fb47f136-7859-4756-bc92-fbcf430b4c8b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.509460] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquired lock "refresh_cache-fb47f136-7859-4756-bc92-fbcf430b4c8b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.509460] env[62552]: DEBUG nova.network.neutron [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.662807] env[62552]: DEBUG nova.network.neutron [req-5379a909-f883-43c3-8237-25588f62dbe8 req-f84c8724-db11-4f40-be1a-4083857a0ca8 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Updated VIF entry in instance network info cache for port c8ca7f73-e9b3-46c6-863c-f304b521dcbc. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.663294] env[62552]: DEBUG nova.network.neutron [req-5379a909-f883-43c3-8237-25588f62dbe8 req-f84c8724-db11-4f40-be1a-4083857a0ca8 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Updating instance_info_cache with network_info: [{"id": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "address": "fa:16:3e:9f:03:3a", "network": {"id": "d102013b-c1eb-4cd4-82f5-c4fba213a4ca", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1492403044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81022b826f084769a961659febdacd4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8ca7f73-e9", "ovs_interfaceid": "c8ca7f73-e9b3-46c6-863c-f304b521dcbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.775297] env[62552]: INFO nova.compute.manager [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Rebuilding instance [ 869.791928] env[62552]: INFO nova.virt.block_device [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Booting with volume 250ad8c9-818e-4ebf-864a-a2e15cd300d3 at /dev/sdb [ 869.830864] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-786d099b-62a6-461f-bba4-e4779ecf4838 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.833700] env[62552]: DEBUG nova.compute.manager [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 869.835195] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92222d7-ae2c-4bbc-9ac2-ed5ec7d6b144 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.847592] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfafbb0c-39f9-4112-b050-3d3ce4290076 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.873500] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.126s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.885533] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.474s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.887105] env[62552]: INFO nova.compute.claims [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.889900] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47745b21-7516-4595-9b07-43416b421862 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.897248] env[62552]: INFO nova.scheduler.client.report [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted allocations for instance 3b333977-0b4b-4ccf-bed5-231215abae3c [ 869.904618] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7461d0b-77bf-49f4-8383-781f28f0ffd8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.940704] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba18099-4e37-403b-8078-a8b6ef70a6d7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.950756] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea21ed48-0b20-4340-9180-6890ea82e9d4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.965354] env[62552]: DEBUG nova.virt.block_device [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating existing volume attachment record: 1de62591-fd4b-4df0-86f6-b62bde86bbc9 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 870.066331] env[62552]: DEBUG nova.network.neutron [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.114814] env[62552]: DEBUG oslo_vmware.rw_handles [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b591c7-af48-100d-415e-a3a0906e6d1c/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 870.115410] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a303724-0bdb-48c0-bd77-ae6cfc8cd51d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.124272] env[62552]: DEBUG oslo_vmware.rw_handles [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b591c7-af48-100d-415e-a3a0906e6d1c/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 870.124402] env[62552]: ERROR oslo_vmware.rw_handles [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b591c7-af48-100d-415e-a3a0906e6d1c/disk-0.vmdk due to incomplete transfer. [ 870.124635] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1c0cf379-6848-4036-a8b3-4d39c1e0892f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.137220] env[62552]: DEBUG oslo_vmware.rw_handles [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b591c7-af48-100d-415e-a3a0906e6d1c/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 870.137436] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Uploaded image 967e9aaa-56d6-402c-abae-6f8af6ce99e2 to the Glance image server {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 870.139818] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 870.140371] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-aad612d0-8366-4f21-ae9a-708e442fc6ea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.146344] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 870.146344] env[62552]: value = "task-1239547" [ 870.146344] env[62552]: _type = "Task" [ 870.146344] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.154110] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239547, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.165916] env[62552]: DEBUG oslo_concurrency.lockutils [req-5379a909-f883-43c3-8237-25588f62dbe8 req-f84c8724-db11-4f40-be1a-4083857a0ca8 service nova] Releasing lock "refresh_cache-1cfeedac-f71e-42e4-a04f-8a0462c85907" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.332753] env[62552]: DEBUG nova.network.neutron [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Updating instance_info_cache with network_info: [{"id": "e421f4d6-d892-4c96-8c11-32f1b7553780", "address": "fa:16:3e:cf:f5:79", "network": {"id": "eb751d71-707e-4572-a811-aba646a77116", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-858940822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51c2a67f74dc42d2b915e0429199f89b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape421f4d6-d8", "ovs_interfaceid": "e421f4d6-d892-4c96-8c11-32f1b7553780", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.420284] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b96c9ad8-cbec-4ed5-a444-d0fbd322b889 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "3b333977-0b4b-4ccf-bed5-231215abae3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.465s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.658694] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239547, 'name': Destroy_Task, 'duration_secs': 0.336881} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.658694] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Destroyed the VM [ 870.658694] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 870.658694] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ff775704-6d89-4d99-ad11-d0a027c2f688 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.664647] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 870.664647] env[62552]: value = "task-1239551" [ 870.664647] env[62552]: _type = "Task" [ 870.664647] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.677944] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239551, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.762584] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updating instance_info_cache with network_info: [{"id": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "address": "fa:16:3e:82:10:f7", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabe02bcf-41", "ovs_interfaceid": "abe02bcf-41b7-4884-8ef3-70b5268480f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.834807] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Releasing lock "refresh_cache-fb47f136-7859-4756-bc92-fbcf430b4c8b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.835172] env[62552]: DEBUG nova.compute.manager [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Instance network_info: |[{"id": "e421f4d6-d892-4c96-8c11-32f1b7553780", "address": "fa:16:3e:cf:f5:79", "network": {"id": "eb751d71-707e-4572-a811-aba646a77116", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-858940822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51c2a67f74dc42d2b915e0429199f89b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape421f4d6-d8", "ovs_interfaceid": "e421f4d6-d892-4c96-8c11-32f1b7553780", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 870.835774] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:f5:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cf748a8-7ae0-4dca-817d-e727c30d72f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e421f4d6-d892-4c96-8c11-32f1b7553780', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.846608] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Creating folder: Project (51c2a67f74dc42d2b915e0429199f89b). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 870.847314] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1dc5ddb9-eff5-4c87-b207-6b9fae02b316 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.861585] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Created folder: Project (51c2a67f74dc42d2b915e0429199f89b) in parent group-v267339. [ 870.862008] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Creating folder: Instances. Parent ref: group-v267442. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 870.862358] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cb39f9b5-21b1-4dc9-9066-9d7c7f29e972 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.866094] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.866791] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a11f57eb-feae-40d0-8fe3-6978226a6e33 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.871668] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Created folder: Instances in parent group-v267442. [ 870.871668] env[62552]: DEBUG oslo.service.loopingcall [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.871952] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.872177] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc460b23-89ae-454a-864e-8eb6158c09d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.893280] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 870.893280] env[62552]: value = "task-1239554" [ 870.893280] env[62552]: _type = "Task" [ 870.893280] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.903330] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.903330] env[62552]: value = "task-1239555" [ 870.903330] env[62552]: _type = "Task" [ 870.903330] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.914614] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239554, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.921383] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239555, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.949526] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.949773] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.988682] env[62552]: DEBUG nova.compute.manager [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Received event network-vif-plugged-e421f4d6-d892-4c96-8c11-32f1b7553780 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 870.989075] env[62552]: DEBUG oslo_concurrency.lockutils [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] Acquiring lock "fb47f136-7859-4756-bc92-fbcf430b4c8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.989333] env[62552]: DEBUG oslo_concurrency.lockutils [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] Lock "fb47f136-7859-4756-bc92-fbcf430b4c8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.989423] env[62552]: DEBUG oslo_concurrency.lockutils [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] Lock "fb47f136-7859-4756-bc92-fbcf430b4c8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.989676] env[62552]: DEBUG nova.compute.manager [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] No waiting events found dispatching network-vif-plugged-e421f4d6-d892-4c96-8c11-32f1b7553780 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 870.989775] env[62552]: WARNING nova.compute.manager [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Received unexpected event network-vif-plugged-e421f4d6-d892-4c96-8c11-32f1b7553780 for instance with vm_state building and task_state spawning. [ 870.989912] env[62552]: DEBUG nova.compute.manager [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Received event network-changed-e421f4d6-d892-4c96-8c11-32f1b7553780 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 870.990084] env[62552]: DEBUG nova.compute.manager [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Refreshing instance network info cache due to event network-changed-e421f4d6-d892-4c96-8c11-32f1b7553780. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 870.990281] env[62552]: DEBUG oslo_concurrency.lockutils [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] Acquiring lock "refresh_cache-fb47f136-7859-4756-bc92-fbcf430b4c8b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.990435] env[62552]: DEBUG oslo_concurrency.lockutils [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] Acquired lock "refresh_cache-fb47f136-7859-4756-bc92-fbcf430b4c8b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.990598] env[62552]: DEBUG nova.network.neutron [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Refreshing network info cache for port e421f4d6-d892-4c96-8c11-32f1b7553780 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.092938] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquiring lock "1cfeedac-f71e-42e4-a04f-8a0462c85907" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.093291] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "1cfeedac-f71e-42e4-a04f-8a0462c85907" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.093551] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquiring lock "1cfeedac-f71e-42e4-a04f-8a0462c85907-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.093794] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "1cfeedac-f71e-42e4-a04f-8a0462c85907-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.094076] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "1cfeedac-f71e-42e4-a04f-8a0462c85907-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.097391] env[62552]: INFO nova.compute.manager [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Terminating instance [ 871.174066] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239551, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.264368] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Releasing lock "refresh_cache-0418260a-aa27-4955-ab15-b180ec04f0b3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.264608] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updated the network info_cache for instance {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 871.264805] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.264987] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.265167] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.265431] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.265604] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.265774] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.265908] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 871.266098] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.291602] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26be7bd-a473-49c8-9ef1-5bec653aa054 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.299797] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645f8b57-914f-44ef-b6eb-15df5f0761e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.332024] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6d3b77-f630-4477-89e3-8ec29f36ce93 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.339996] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b448c97a-ec7d-47ae-8001-8e6b9004894a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.354086] env[62552]: DEBUG nova.compute.provider_tree [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.404037] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239554, 'name': PowerOffVM_Task, 'duration_secs': 0.152537} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.404365] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 871.404635] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.406629] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6a4681-dd82-4d68-a051-c2a078464172 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.416474] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239555, 'name': CreateVM_Task, 'duration_secs': 0.38247} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.418352] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.418642] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 871.419371] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.420031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.420031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.420153] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1fc18891-6ff9-467b-9836-f58f1b487956 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.421987] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18525600-4f1b-40b8-898b-beeb8483536d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.426271] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 871.426271] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523c4abe-6134-bddd-5c90-b47732876c4f" [ 871.426271] env[62552]: _type = "Task" [ 871.426271] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.433857] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523c4abe-6134-bddd-5c90-b47732876c4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.445587] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 871.445821] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 871.446018] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Deleting the datastore file [datastore1] fea83dde-3181-49dd-a000-5e5cffc8de95 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 871.446282] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c115fe8-08ad-46dc-8310-0190e39efd6b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.452221] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 871.452221] env[62552]: value = "task-1239557" [ 871.452221] env[62552]: _type = "Task" [ 871.452221] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.455678] env[62552]: DEBUG nova.compute.manager [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 871.463720] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239557, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.601580] env[62552]: DEBUG nova.compute.manager [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 871.601803] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.602700] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7d03b6-1e0b-4553-acd4-772756c1a64b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.612618] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.612878] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b4b446f-76df-4564-9f90-666418931b57 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.619597] env[62552]: DEBUG oslo_vmware.api [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 871.619597] env[62552]: value = "task-1239558" [ 871.619597] env[62552]: _type = "Task" [ 871.619597] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.628465] env[62552]: DEBUG oslo_vmware.api [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239558, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.675444] env[62552]: DEBUG oslo_vmware.api [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239551, 'name': RemoveSnapshot_Task, 'duration_secs': 0.875173} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.675717] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 871.675998] env[62552]: INFO nova.compute.manager [None req-d7888ded-166d-4be0-b204-94432ebeceff tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Took 16.65 seconds to snapshot the instance on the hypervisor. [ 871.753556] env[62552]: DEBUG nova.network.neutron [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Updated VIF entry in instance network info cache for port e421f4d6-d892-4c96-8c11-32f1b7553780. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.753956] env[62552]: DEBUG nova.network.neutron [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Updating instance_info_cache with network_info: [{"id": "e421f4d6-d892-4c96-8c11-32f1b7553780", "address": "fa:16:3e:cf:f5:79", "network": {"id": "eb751d71-707e-4572-a811-aba646a77116", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-858940822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51c2a67f74dc42d2b915e0429199f89b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape421f4d6-d8", "ovs_interfaceid": "e421f4d6-d892-4c96-8c11-32f1b7553780", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.768977] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.857848] env[62552]: DEBUG nova.scheduler.client.report [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 871.936446] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523c4abe-6134-bddd-5c90-b47732876c4f, 'name': SearchDatastore_Task, 'duration_secs': 0.031506} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.938024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.938024] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.938024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.938024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.938024] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.938024] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3de9e17-2bae-4bfb-a0a9-bc368172f42f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.945390] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.945563] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.946245] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0c58cac-78d6-4b5b-aa21-ceceefaefdd5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.951313] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 871.951313] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b5e4ca-d8a2-276b-4720-44cc38a234ab" [ 871.951313] env[62552]: _type = "Task" [ 871.951313] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.964028] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b5e4ca-d8a2-276b-4720-44cc38a234ab, 'name': SearchDatastore_Task, 'duration_secs': 0.007793} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.969690] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239557, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135509} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.970137] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ba102d7-8c08-4609-b203-b9a5e2093488 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.972183] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.972380] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.972559] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.977646] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 871.977646] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525f900c-8d0a-919b-99d3-f66eb1b4e46b" [ 871.977646] env[62552]: _type = "Task" [ 871.977646] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.982743] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.985600] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525f900c-8d0a-919b-99d3-f66eb1b4e46b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.128464] env[62552]: DEBUG oslo_vmware.api [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239558, 'name': PowerOffVM_Task, 'duration_secs': 0.19664} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.128902] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.128957] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.129305] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d07309e9-9d24-4e59-a4f8-9e550bc70406 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.192838] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.193141] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.193340] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Deleting the datastore file [datastore2] 1cfeedac-f71e-42e4-a04f-8a0462c85907 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.193601] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f6036cf-ce56-4dfb-a1f9-7ae8198c2198 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.200573] env[62552]: DEBUG oslo_vmware.api [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for the task: (returnval){ [ 872.200573] env[62552]: value = "task-1239560" [ 872.200573] env[62552]: _type = "Task" [ 872.200573] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.210256] env[62552]: DEBUG oslo_vmware.api [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239560, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.257844] env[62552]: DEBUG oslo_concurrency.lockutils [req-964bcadc-5e6f-4eb8-880a-1011aa672bbc req-7e8ee856-b565-4f2d-b42f-5cba646380b3 service nova] Releasing lock "refresh_cache-fb47f136-7859-4756-bc92-fbcf430b4c8b" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.364030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.364030] env[62552]: DEBUG nova.compute.manager [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 872.368686] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.311s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.372151] env[62552]: INFO nova.compute.claims [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.490119] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525f900c-8d0a-919b-99d3-f66eb1b4e46b, 'name': SearchDatastore_Task, 'duration_secs': 0.007974} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.490423] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.491033] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] fb47f136-7859-4756-bc92-fbcf430b4c8b/fb47f136-7859-4756-bc92-fbcf430b4c8b.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.491033] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3bca8e8b-35f7-49ce-80ad-a09bf964af75 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.497066] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 872.497066] env[62552]: value = "task-1239562" [ 872.497066] env[62552]: _type = "Task" [ 872.497066] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.505082] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239562, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.629803] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 872.629803] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267439', 'volume_id': 'f925d139-4314-4651-be8c-f07a21f94eea', 'name': 'volume-f925d139-4314-4651-be8c-f07a21f94eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80478878-ff82-4ed6-a851-8eb2bec01e22', 'attached_at': '', 'detached_at': '', 'volume_id': 'f925d139-4314-4651-be8c-f07a21f94eea', 'serial': 'f925d139-4314-4651-be8c-f07a21f94eea'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 872.630780] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5de20fb-a8fb-4879-ad8e-ea63da34ae66 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.651334] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb19ecc4-164f-440f-a294-3fcd3d6634cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.678624] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] volume-f925d139-4314-4651-be8c-f07a21f94eea/volume-f925d139-4314-4651-be8c-f07a21f94eea.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.678848] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-064f27c3-4648-4f4b-ae97-76eb4ef05fc3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.699246] env[62552]: DEBUG oslo_vmware.api [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 872.699246] env[62552]: value = "task-1239563" [ 872.699246] env[62552]: _type = "Task" [ 872.699246] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.712257] env[62552]: DEBUG oslo_vmware.api [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Task: {'id': task-1239560, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134742} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.716618] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.716618] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 872.716765] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 872.716915] env[62552]: INFO nova.compute.manager [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Took 1.12 seconds to destroy the instance on the hypervisor. [ 872.717194] env[62552]: DEBUG oslo.service.loopingcall [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.717416] env[62552]: DEBUG oslo_vmware.api [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239563, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.717683] env[62552]: DEBUG nova.compute.manager [-] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 872.717832] env[62552]: DEBUG nova.network.neutron [-] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 872.879962] env[62552]: DEBUG nova.compute.utils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.881958] env[62552]: DEBUG nova.compute.manager [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 872.882389] env[62552]: DEBUG nova.network.neutron [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 873.009447] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239562, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483598} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.009768] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] fb47f136-7859-4756-bc92-fbcf430b4c8b/fb47f136-7859-4756-bc92-fbcf430b4c8b.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 873.010044] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.010533] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7ad37ec-ca7e-4175-9bca-781466af1719 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.014853] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.015130] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.015356] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.015765] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.015965] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.016216] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.016517] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.016718] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.017948] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.018077] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.018315] env[62552]: DEBUG nova.virt.hardware [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.019220] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6cd12f-53bd-40b5-a703-ee48c33d23e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.031593] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0876c2f-4b93-4951-a8fc-d42a76264c2f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.036822] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 873.036822] env[62552]: value = "task-1239564" [ 873.036822] env[62552]: _type = "Task" [ 873.036822] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.053036] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.059059] env[62552]: DEBUG oslo.service.loopingcall [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.060024] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 873.060343] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-592eb0f0-bb5d-4754-9aa5-4348ad524f64 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.078213] env[62552]: DEBUG nova.policy [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d904ba8cf284314baf512c7ea33554b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1716b64900f44a8fb2282801905efd67', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.080149] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239564, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.084674] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.084674] env[62552]: value = "task-1239565" [ 873.084674] env[62552]: _type = "Task" [ 873.084674] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.094785] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239565, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.211704] env[62552]: DEBUG oslo_vmware.api [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239563, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.360731] env[62552]: DEBUG nova.compute.manager [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 873.361796] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc9e7d6f-2f22-472f-b2d4-3eebcb3d1f56 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.387286] env[62552]: DEBUG nova.compute.manager [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 873.550336] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239564, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058499} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.550697] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.551529] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637a6c33-8090-4eff-85a0-850fe71802d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.579158] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] fb47f136-7859-4756-bc92-fbcf430b4c8b/fb47f136-7859-4756-bc92-fbcf430b4c8b.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.582709] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e02e3e7-6fee-4e6e-aabd-a9e176a56438 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.615461] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239565, 'name': CreateVM_Task, 'duration_secs': 0.353963} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.617438] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.618132] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 873.618132] env[62552]: value = "task-1239566" [ 873.618132] env[62552]: _type = "Task" [ 873.618132] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.618587] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.618786] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.619174] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.619555] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52fb8eef-7094-4aa8-a547-dd57b4fe5d81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.630585] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 873.630585] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525c7a3a-54b7-df8c-bd54-e1e004e449a9" [ 873.630585] env[62552]: _type = "Task" [ 873.630585] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.634354] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239566, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.648105] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525c7a3a-54b7-df8c-bd54-e1e004e449a9, 'name': SearchDatastore_Task, 'duration_secs': 0.011354} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.648459] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.648725] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.648981] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.649159] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.649397] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.649681] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fba47061-0318-40bd-90ce-ea0246a3af42 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.661070] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.661328] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.662166] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97ff626d-388f-4b7f-95ed-bd5e201bce14 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.670524] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 873.670524] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525af4d9-44b5-0258-85e5-7de8d460a38c" [ 873.670524] env[62552]: _type = "Task" [ 873.670524] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.678541] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525af4d9-44b5-0258-85e5-7de8d460a38c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.709555] env[62552]: DEBUG oslo_vmware.api [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239563, 'name': ReconfigVM_Task, 'duration_secs': 0.705394} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.709978] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Reconfigured VM instance instance-0000003a to attach disk [datastore1] volume-f925d139-4314-4651-be8c-f07a21f94eea/volume-f925d139-4314-4651-be8c-f07a21f94eea.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.717114] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec71e0dd-941e-4d68-8df9-896d108b9295 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.733093] env[62552]: DEBUG oslo_vmware.api [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 873.733093] env[62552]: value = "task-1239567" [ 873.733093] env[62552]: _type = "Task" [ 873.733093] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.743478] env[62552]: DEBUG oslo_vmware.api [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239567, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.768358] env[62552]: DEBUG nova.network.neutron [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Successfully created port: cf27f46e-a6d0-4547-9eea-2a88fb1f21dc {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.875478] env[62552]: INFO nova.compute.manager [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] instance snapshotting [ 873.876434] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c376ed-7e36-4946-a029-52604013c58c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.907373] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66cb62e-bb43-4a00-8b10-0a757187524b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.917826] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b660123f-d87c-4cfe-b86e-528f16257cd0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.925602] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ebe5e5-c868-4959-aa91-c378b69198c1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.963479] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0621882a-4e84-4f38-afe7-28ffba5f4739 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.971620] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5c45b4-8cb7-4134-8bc9-9f8b298569f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.985409] env[62552]: DEBUG nova.compute.provider_tree [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.107083] env[62552]: DEBUG nova.network.neutron [-] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.129876] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239566, 'name': ReconfigVM_Task, 'duration_secs': 0.376955} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.130184] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Reconfigured VM instance instance-00000046 to attach disk [datastore1] fb47f136-7859-4756-bc92-fbcf430b4c8b/fb47f136-7859-4756-bc92-fbcf430b4c8b.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.130880] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6a1373b-7f38-40f0-8911-dd5958dd4b3f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.137018] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 874.137018] env[62552]: value = "task-1239568" [ 874.137018] env[62552]: _type = "Task" [ 874.137018] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.146615] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239568, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.180474] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525af4d9-44b5-0258-85e5-7de8d460a38c, 'name': SearchDatastore_Task, 'duration_secs': 0.00977} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.181393] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03c02e1d-ff78-4511-ba2a-804457b5cb7b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.189101] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 874.189101] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52072cfd-a365-549d-8a64-ae525ebb98bd" [ 874.189101] env[62552]: _type = "Task" [ 874.189101] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.194575] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52072cfd-a365-549d-8a64-ae525ebb98bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.242956] env[62552]: DEBUG oslo_vmware.api [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239567, 'name': ReconfigVM_Task, 'duration_secs': 0.144519} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.243316] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267439', 'volume_id': 'f925d139-4314-4651-be8c-f07a21f94eea', 'name': 'volume-f925d139-4314-4651-be8c-f07a21f94eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80478878-ff82-4ed6-a851-8eb2bec01e22', 'attached_at': '', 'detached_at': '', 'volume_id': 'f925d139-4314-4651-be8c-f07a21f94eea', 'serial': 'f925d139-4314-4651-be8c-f07a21f94eea'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 874.274640] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fcbe393c-7711-483d-bc96-718acb67cc4a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "7d43e8ac-aa40-4758-8ca9-1e0e2c6791e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.274892] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fcbe393c-7711-483d-bc96-718acb67cc4a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "7d43e8ac-aa40-4758-8ca9-1e0e2c6791e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.314668] env[62552]: DEBUG nova.compute.manager [req-47d348fb-5aa6-402d-836e-b6fd1132af53 req-85db2522-74f6-4b19-b858-fdecfc6b1681 service nova] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Received event network-vif-deleted-c8ca7f73-e9b3-46c6-863c-f304b521dcbc {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 874.411929] env[62552]: DEBUG nova.compute.manager [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 874.421988] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 874.422319] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d8759540-5f53-4f80-b527-e3afecdc9827 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.430897] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 874.430897] env[62552]: value = "task-1239569" [ 874.430897] env[62552]: _type = "Task" [ 874.430897] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.440251] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.440513] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.440678] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.440887] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.441081] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.441239] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.441450] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.441613] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.441782] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.441953] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.442201] env[62552]: DEBUG nova.virt.hardware [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.442941] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f5b1ce-9d85-4ad7-a4b0-98cfb9e790b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.448440] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239569, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.453304] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d5e908-7715-4861-8172-5698db34115d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.487840] env[62552]: DEBUG nova.scheduler.client.report [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 874.609565] env[62552]: INFO nova.compute.manager [-] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Took 1.89 seconds to deallocate network for instance. [ 874.649011] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239568, 'name': Rename_Task, 'duration_secs': 0.139356} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.649011] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 874.649011] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09dd48db-6941-4863-aae8-e7dce3c2546c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.655050] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 874.655050] env[62552]: value = "task-1239570" [ 874.655050] env[62552]: _type = "Task" [ 874.655050] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.663141] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.698151] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52072cfd-a365-549d-8a64-ae525ebb98bd, 'name': SearchDatastore_Task, 'duration_secs': 0.009854} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.698485] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.698852] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] fea83dde-3181-49dd-a000-5e5cffc8de95/fea83dde-3181-49dd-a000-5e5cffc8de95.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.699148] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f27bbca-c871-48bf-a43d-60ec373b18e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.705547] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 874.705547] env[62552]: value = "task-1239571" [ 874.705547] env[62552]: _type = "Task" [ 874.705547] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.713362] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.941692] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239569, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.993157] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.624s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.993858] env[62552]: DEBUG nova.compute.manager [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 874.996765] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.900s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.997017] env[62552]: DEBUG nova.objects.instance [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lazy-loading 'resources' on Instance uuid df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.116590] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.165643] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239570, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.217318] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239571, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.282151] env[62552]: DEBUG nova.objects.instance [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lazy-loading 'flavor' on Instance uuid 80478878-ff82-4ed6-a851-8eb2bec01e22 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.442997] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239569, 'name': CreateSnapshot_Task, 'duration_secs': 0.928192} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.443549] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 875.444682] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d71cf9b-015e-407c-a1d4-9a9d795a2fe8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.499905] env[62552]: DEBUG nova.compute.utils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.504643] env[62552]: DEBUG nova.compute.manager [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 875.504916] env[62552]: DEBUG nova.network.neutron [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.589703] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.625423] env[62552]: DEBUG nova.policy [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '384d6877b7384d21b12a155d13056a7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f13a11803d1b44178f053cc0a4affedd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 875.674841] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239570, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.717848] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589847} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.718177] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] fea83dde-3181-49dd-a000-5e5cffc8de95/fea83dde-3181-49dd-a000-5e5cffc8de95.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.718396] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.718679] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cf15c0b-f8d0-45fc-a093-63b82a0fbfdf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.728059] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 875.728059] env[62552]: value = "task-1239572" [ 875.728059] env[62552]: _type = "Task" [ 875.728059] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.741732] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239572, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.790689] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d4c15c72-4c88-4775-a099-590b747d72ce tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.789s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.824434] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.824804] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.894680] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.895180] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.895431] env[62552]: DEBUG nova.compute.manager [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 875.896997] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7c3a5c-ec5f-48bd-babe-e41d9157432b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.904599] env[62552]: DEBUG nova.compute.manager [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62552) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 875.905632] env[62552]: DEBUG nova.objects.instance [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'flavor' on Instance uuid 88d0f502-fc3f-429e-bdf6-a1ebacec2117 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.962742] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 875.964369] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-53be20f9-7370-49bf-bada-d8e629e62acd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.967774] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1863ae37-b42f-4727-af34-355d6bb2a40f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.976438] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35016bf9-7c55-4b6e-af5d-529dfa23231f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.979732] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 875.979732] env[62552]: value = "task-1239573" [ 875.979732] env[62552]: _type = "Task" [ 875.979732] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.015637] env[62552]: DEBUG nova.compute.manager [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 876.023494] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df01487-3448-4bb3-a9d7-c60cf5282c39 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.032859] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239573, 'name': CloneVM_Task} progress is 23%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.038603] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0619be5-3fa9-4199-8ae0-9d18be372eb9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.054011] env[62552]: DEBUG nova.compute.provider_tree [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.086660] env[62552]: DEBUG nova.network.neutron [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Successfully updated port: cf27f46e-a6d0-4547-9eea-2a88fb1f21dc {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.169930] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239570, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.238314] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239572, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150024} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.238670] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.239510] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9f2c56-6bfa-4ebb-a36d-62889ffce4ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.261692] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] fea83dde-3181-49dd-a000-5e5cffc8de95/fea83dde-3181-49dd-a000-5e5cffc8de95.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.262145] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f7eab9a-f061-4fe7-8d07-930852895a0e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.277944] env[62552]: DEBUG nova.network.neutron [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Successfully created port: e28a5083-5362-482d-a79c-f2103aec3b9f {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.285398] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 876.285398] env[62552]: value = "task-1239574" [ 876.285398] env[62552]: _type = "Task" [ 876.285398] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.294245] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239574, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.406374] env[62552]: DEBUG nova.compute.manager [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Received event network-vif-plugged-cf27f46e-a6d0-4547-9eea-2a88fb1f21dc {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 876.406597] env[62552]: DEBUG oslo_concurrency.lockutils [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] Acquiring lock "5cee4242-49ae-4cb7-a208-e2982f52fbad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.406806] env[62552]: DEBUG oslo_concurrency.lockutils [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] Lock "5cee4242-49ae-4cb7-a208-e2982f52fbad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.406982] env[62552]: DEBUG oslo_concurrency.lockutils [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] Lock "5cee4242-49ae-4cb7-a208-e2982f52fbad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.407166] env[62552]: DEBUG nova.compute.manager [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] No waiting events found dispatching network-vif-plugged-cf27f46e-a6d0-4547-9eea-2a88fb1f21dc {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 876.407338] env[62552]: WARNING nova.compute.manager [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Received unexpected event network-vif-plugged-cf27f46e-a6d0-4547-9eea-2a88fb1f21dc for instance with vm_state building and task_state spawning. [ 876.407502] env[62552]: DEBUG nova.compute.manager [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Received event network-changed-cf27f46e-a6d0-4547-9eea-2a88fb1f21dc {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 876.407660] env[62552]: DEBUG nova.compute.manager [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Refreshing instance network info cache due to event network-changed-cf27f46e-a6d0-4547-9eea-2a88fb1f21dc. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 876.408177] env[62552]: DEBUG oslo_concurrency.lockutils [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] Acquiring lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.408370] env[62552]: DEBUG oslo_concurrency.lockutils [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] Acquired lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.408545] env[62552]: DEBUG nova.network.neutron [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Refreshing network info cache for port cf27f46e-a6d0-4547-9eea-2a88fb1f21dc {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.489147] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239573, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.557216] env[62552]: DEBUG nova.scheduler.client.report [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 876.589468] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.669275] env[62552]: DEBUG oslo_vmware.api [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239570, 'name': PowerOnVM_Task, 'duration_secs': 1.654193} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.673124] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 876.673124] env[62552]: INFO nova.compute.manager [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Took 9.55 seconds to spawn the instance on the hypervisor. [ 876.673124] env[62552]: DEBUG nova.compute.manager [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 876.673124] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d16488-2ae3-43f3-8fbe-07db5e937dfc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.799646] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239574, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.858291] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "80478878-ff82-4ed6-a851-8eb2bec01e22" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.861750] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.917521] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.918097] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39b59998-f735-439f-835a-67b169130af1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.926067] env[62552]: DEBUG oslo_vmware.api [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 876.926067] env[62552]: value = "task-1239575" [ 876.926067] env[62552]: _type = "Task" [ 876.926067] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.935017] env[62552]: DEBUG oslo_vmware.api [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239575, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.958061] env[62552]: DEBUG nova.network.neutron [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.989908] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239573, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.035940] env[62552]: DEBUG nova.compute.manager [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 877.063030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.066s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.065606] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.043s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.067611] env[62552]: INFO nova.compute.claims [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.073061] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='de3d468fc4cbeb053fb8fc7e1da595fa',container_format='bare',created_at=2024-10-10T11:35:09Z,direct_url=,disk_format='vmdk',id=35ac675b-d35f-462e-aaa2-a2061f578780,min_disk=1,min_ram=0,name='tempest-test-snap-1814268988',owner='f13a11803d1b44178f053cc0a4affedd',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-10T11:35:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.073977] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.074825] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.075082] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.077139] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.077259] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.078063] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.078063] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.078063] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.078063] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.078326] env[62552]: DEBUG nova.virt.hardware [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.079177] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8413262a-3df2-4a11-a881-cd50e328e38b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.088987] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1473178-11b7-4f1c-a7bc-4507183a5afc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.105547] env[62552]: INFO nova.scheduler.client.report [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Deleted allocations for instance df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28 [ 877.107382] env[62552]: DEBUG nova.network.neutron [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.191013] env[62552]: INFO nova.compute.manager [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Took 41.15 seconds to build instance. [ 877.295832] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239574, 'name': ReconfigVM_Task, 'duration_secs': 0.671516} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.296130] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Reconfigured VM instance instance-00000045 to attach disk [datastore2] fea83dde-3181-49dd-a000-5e5cffc8de95/fea83dde-3181-49dd-a000-5e5cffc8de95.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.296782] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f73f966-42a3-4f81-8504-f778e27937eb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.306996] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 877.306996] env[62552]: value = "task-1239576" [ 877.306996] env[62552]: _type = "Task" [ 877.306996] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.317200] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239576, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.364142] env[62552]: INFO nova.compute.manager [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Detaching volume f925d139-4314-4651-be8c-f07a21f94eea [ 877.402996] env[62552]: INFO nova.virt.block_device [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Attempting to driver detach volume f925d139-4314-4651-be8c-f07a21f94eea from mountpoint /dev/sdb [ 877.403355] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 877.403596] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267439', 'volume_id': 'f925d139-4314-4651-be8c-f07a21f94eea', 'name': 'volume-f925d139-4314-4651-be8c-f07a21f94eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80478878-ff82-4ed6-a851-8eb2bec01e22', 'attached_at': '', 'detached_at': '', 'volume_id': 'f925d139-4314-4651-be8c-f07a21f94eea', 'serial': 'f925d139-4314-4651-be8c-f07a21f94eea'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 877.404610] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643204f6-7930-4a1f-81be-9d922be15a66 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.431950] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca87c27-b04d-416e-8c95-4ff9ef4ec7f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.439838] env[62552]: DEBUG oslo_vmware.api [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239575, 'name': PowerOffVM_Task, 'duration_secs': 0.201437} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.442142] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 877.442142] env[62552]: DEBUG nova.compute.manager [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 877.443498] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8569111d-16f1-4f96-a198-e962af2f5dd2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.447241] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53815ff2-4204-40f3-a9ee-d1a1c8197746 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.473671] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692098aa-fcb3-4d71-83dc-d3e7b57b979c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.491055] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] The volume has not been displaced from its original location: [datastore1] volume-f925d139-4314-4651-be8c-f07a21f94eea/volume-f925d139-4314-4651-be8c-f07a21f94eea.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 877.496338] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Reconfiguring VM instance instance-0000003a to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 877.501308] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9dfe95b2-b231-4232-b876-8090944f2f21 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.522550] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239573, 'name': CloneVM_Task, 'duration_secs': 1.420786} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.524203] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Created linked-clone VM from snapshot [ 877.524578] env[62552]: DEBUG oslo_vmware.api [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 877.524578] env[62552]: value = "task-1239577" [ 877.524578] env[62552]: _type = "Task" [ 877.524578] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.525374] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e88b9e-ae0b-4daf-a397-30257a14edd0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.537439] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Uploading image fd4b6cf6-b251-4731-93a3-ed4f4acab88b {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 877.543292] env[62552]: DEBUG oslo_vmware.api [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239577, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.563771] env[62552]: DEBUG oslo_vmware.rw_handles [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 877.563771] env[62552]: value = "vm-267447" [ 877.563771] env[62552]: _type = "VirtualMachine" [ 877.563771] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 877.564119] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-25c900a7-cb5b-43dc-ad6b-47bac30fb86d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.572827] env[62552]: DEBUG oslo_vmware.rw_handles [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lease: (returnval){ [ 877.572827] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1313f-0913-d485-66f5-230ad0d3fffd" [ 877.572827] env[62552]: _type = "HttpNfcLease" [ 877.572827] env[62552]: } obtained for exporting VM: (result){ [ 877.572827] env[62552]: value = "vm-267447" [ 877.572827] env[62552]: _type = "VirtualMachine" [ 877.572827] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 877.573091] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the lease: (returnval){ [ 877.573091] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1313f-0913-d485-66f5-230ad0d3fffd" [ 877.573091] env[62552]: _type = "HttpNfcLease" [ 877.573091] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 877.582462] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.582462] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1313f-0913-d485-66f5-230ad0d3fffd" [ 877.582462] env[62552]: _type = "HttpNfcLease" [ 877.582462] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 877.612901] env[62552]: DEBUG oslo_concurrency.lockutils [req-33c22783-5eab-470b-9154-b8f91aad1bdb req-13c49d8a-d896-4c1b-b701-a1750dc08e30 service nova] Releasing lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.615195] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.615377] env[62552]: DEBUG nova.network.neutron [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.616508] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0d3efad7-8ae4-4534-bc84-a4ca7819c6cc tempest-SecurityGroupsTestJSON-789577123 tempest-SecurityGroupsTestJSON-789577123-project-member] Lock "df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.132s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.695034] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bc859928-17fe-419e-a14a-b8405de75fcd tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "fb47f136-7859-4756-bc92-fbcf430b4c8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.831s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.808458] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquiring lock "fb47f136-7859-4756-bc92-fbcf430b4c8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.808756] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "fb47f136-7859-4756-bc92-fbcf430b4c8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.808974] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquiring lock "fb47f136-7859-4756-bc92-fbcf430b4c8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.809190] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "fb47f136-7859-4756-bc92-fbcf430b4c8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.809367] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "fb47f136-7859-4756-bc92-fbcf430b4c8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.813854] env[62552]: INFO nova.compute.manager [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Terminating instance [ 877.824808] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239576, 'name': Rename_Task, 'duration_secs': 0.258409} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.825134] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 877.825404] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80368574-3d50-4f29-81d6-514811aed992 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.833354] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 877.833354] env[62552]: value = "task-1239579" [ 877.833354] env[62552]: _type = "Task" [ 877.833354] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.843618] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239579, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.977893] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cdcbeeba-7ced-41e9-81be-022a2a522d0e tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.083s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.038595] env[62552]: DEBUG oslo_vmware.api [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239577, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.085808] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 878.085808] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1313f-0913-d485-66f5-230ad0d3fffd" [ 878.085808] env[62552]: _type = "HttpNfcLease" [ 878.085808] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 878.086117] env[62552]: DEBUG oslo_vmware.rw_handles [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 878.086117] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e1313f-0913-d485-66f5-230ad0d3fffd" [ 878.086117] env[62552]: _type = "HttpNfcLease" [ 878.086117] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 878.086931] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21888967-331f-4f3e-a7d4-fbb5b70fbfdd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.097993] env[62552]: DEBUG oslo_vmware.rw_handles [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527499e9-eb32-4eef-49a1-e96bfcd355eb/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 878.098212] env[62552]: DEBUG oslo_vmware.rw_handles [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527499e9-eb32-4eef-49a1-e96bfcd355eb/disk-0.vmdk for reading. {{(pid=62552) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 878.199623] env[62552]: DEBUG nova.compute.manager [None req-fcbe393c-7711-483d-bc96-718acb67cc4a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 7d43e8ac-aa40-4758-8ca9-1e0e2c6791e1] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 878.243989] env[62552]: DEBUG nova.network.neutron [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.285720] env[62552]: DEBUG nova.network.neutron [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Successfully updated port: e28a5083-5362-482d-a79c-f2103aec3b9f {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.313553] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-48c1be51-2efe-46a2-a342-99b49fafb5e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.320282] env[62552]: DEBUG nova.compute.manager [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 878.320517] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 878.321546] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22ed075-64a7-46a7-b044-133051cb3bf7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.329781] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 878.332438] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f029511-7da7-4142-9b66-c9e57c425915 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.345058] env[62552]: DEBUG oslo_vmware.api [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 878.345058] env[62552]: value = "task-1239580" [ 878.345058] env[62552]: _type = "Task" [ 878.345058] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.348709] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239579, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.361687] env[62552]: DEBUG oslo_vmware.api [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239580, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.518109] env[62552]: DEBUG nova.network.neutron [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Updating instance_info_cache with network_info: [{"id": "cf27f46e-a6d0-4547-9eea-2a88fb1f21dc", "address": "fa:16:3e:04:e5:1a", "network": {"id": "925fdab2-f263-46f3-ab17-2f51407bae3e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1178354148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1716b64900f44a8fb2282801905efd67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf27f46e-a6", "ovs_interfaceid": "cf27f46e-a6d0-4547-9eea-2a88fb1f21dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.551070] env[62552]: DEBUG oslo_vmware.api [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239577, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.578652] env[62552]: DEBUG nova.compute.manager [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Received event network-vif-plugged-e28a5083-5362-482d-a79c-f2103aec3b9f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 878.578652] env[62552]: DEBUG oslo_concurrency.lockutils [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] Acquiring lock "526d0f25-3e99-4558-94c0-754ec2a80bad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.578823] env[62552]: DEBUG oslo_concurrency.lockutils [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] Lock "526d0f25-3e99-4558-94c0-754ec2a80bad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.579136] env[62552]: DEBUG oslo_concurrency.lockutils [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] Lock "526d0f25-3e99-4558-94c0-754ec2a80bad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.579434] env[62552]: DEBUG nova.compute.manager [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] No waiting events found dispatching network-vif-plugged-e28a5083-5362-482d-a79c-f2103aec3b9f {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 878.579752] env[62552]: WARNING nova.compute.manager [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Received unexpected event network-vif-plugged-e28a5083-5362-482d-a79c-f2103aec3b9f for instance with vm_state building and task_state spawning. [ 878.580083] env[62552]: DEBUG nova.compute.manager [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Received event network-changed-e28a5083-5362-482d-a79c-f2103aec3b9f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 878.580440] env[62552]: DEBUG nova.compute.manager [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Refreshing instance network info cache due to event network-changed-e28a5083-5362-482d-a79c-f2103aec3b9f. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 878.580790] env[62552]: DEBUG oslo_concurrency.lockutils [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] Acquiring lock "refresh_cache-526d0f25-3e99-4558-94c0-754ec2a80bad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.581146] env[62552]: DEBUG oslo_concurrency.lockutils [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] Acquired lock "refresh_cache-526d0f25-3e99-4558-94c0-754ec2a80bad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.581518] env[62552]: DEBUG nova.network.neutron [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Refreshing network info cache for port e28a5083-5362-482d-a79c-f2103aec3b9f {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.606288] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115794b9-a607-438c-b3bb-b2fd6aab911b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.618759] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9762f81a-1f11-4442-8ef1-0a38ebac8080 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.672082] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3da359-ebdc-4fcc-919e-1e53f7165926 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.676944] env[62552]: DEBUG nova.objects.instance [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'flavor' on Instance uuid 88d0f502-fc3f-429e-bdf6-a1ebacec2117 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.687022] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733beba1-8808-49cf-8ec5-4183a432cb3f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.710319] env[62552]: DEBUG nova.compute.manager [None req-fcbe393c-7711-483d-bc96-718acb67cc4a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 7d43e8ac-aa40-4758-8ca9-1e0e2c6791e1] Instance disappeared before build. {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2445}} [ 878.712184] env[62552]: DEBUG nova.compute.provider_tree [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.789669] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "refresh_cache-526d0f25-3e99-4558-94c0-754ec2a80bad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.847083] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239579, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.858189] env[62552]: DEBUG oslo_vmware.api [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239580, 'name': PowerOffVM_Task, 'duration_secs': 0.302203} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.858481] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 878.859194] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 878.859499] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6925cd7f-1325-4122-aa43-e7f324d4a9a2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.922481] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 878.922729] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 878.922909] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Deleting the datastore file [datastore1] fb47f136-7859-4756-bc92-fbcf430b4c8b {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 878.923238] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f8cfca92-7272-4edb-a042-25f92b351219 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.930697] env[62552]: DEBUG oslo_vmware.api [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for the task: (returnval){ [ 878.930697] env[62552]: value = "task-1239582" [ 878.930697] env[62552]: _type = "Task" [ 878.930697] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.942596] env[62552]: DEBUG oslo_vmware.api [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239582, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.022054] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.022390] env[62552]: DEBUG nova.compute.manager [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Instance network_info: |[{"id": "cf27f46e-a6d0-4547-9eea-2a88fb1f21dc", "address": "fa:16:3e:04:e5:1a", "network": {"id": "925fdab2-f263-46f3-ab17-2f51407bae3e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1178354148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1716b64900f44a8fb2282801905efd67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf27f46e-a6", "ovs_interfaceid": "cf27f46e-a6d0-4547-9eea-2a88fb1f21dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 879.022847] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:e5:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf27f46e-a6d0-4547-9eea-2a88fb1f21dc', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.032587] env[62552]: DEBUG oslo.service.loopingcall [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 879.032990] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.037564] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4184bc19-0300-4179-872a-856b2fde7430 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.062777] env[62552]: DEBUG oslo_vmware.api [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239577, 'name': ReconfigVM_Task, 'duration_secs': 1.290339} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.064507] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Reconfigured VM instance instance-0000003a to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 879.069654] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.069654] env[62552]: value = "task-1239583" [ 879.069654] env[62552]: _type = "Task" [ 879.069654] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.070067] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c6a731d-ac15-4440-af41-fd40938e3321 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.093098] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239583, 'name': CreateVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.095079] env[62552]: DEBUG oslo_vmware.api [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 879.095079] env[62552]: value = "task-1239584" [ 879.095079] env[62552]: _type = "Task" [ 879.095079] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.104728] env[62552]: DEBUG oslo_vmware.api [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239584, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.132135] env[62552]: DEBUG nova.network.neutron [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 879.182698] env[62552]: DEBUG oslo_concurrency.lockutils [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.182872] env[62552]: DEBUG oslo_concurrency.lockutils [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.185926] env[62552]: DEBUG nova.network.neutron [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.185926] env[62552]: DEBUG nova.objects.instance [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'info_cache' on Instance uuid 88d0f502-fc3f-429e-bdf6-a1ebacec2117 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.216560] env[62552]: DEBUG nova.scheduler.client.report [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 879.229917] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fcbe393c-7711-483d-bc96-718acb67cc4a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "7d43e8ac-aa40-4758-8ca9-1e0e2c6791e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 4.955s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.262693] env[62552]: DEBUG nova.network.neutron [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.347910] env[62552]: DEBUG oslo_vmware.api [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239579, 'name': PowerOnVM_Task, 'duration_secs': 1.300723} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.348366] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.348654] env[62552]: DEBUG nova.compute.manager [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 879.353506] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759cd4f3-80a4-44d2-84e8-e5b2ede32a6f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.441153] env[62552]: DEBUG oslo_vmware.api [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Task: {'id': task-1239582, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.262217} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.441432] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 879.441718] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 879.442035] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 879.442259] env[62552]: INFO nova.compute.manager [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 879.442563] env[62552]: DEBUG oslo.service.loopingcall [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 879.442818] env[62552]: DEBUG nova.compute.manager [-] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 879.442995] env[62552]: DEBUG nova.network.neutron [-] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 879.590609] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239583, 'name': CreateVM_Task, 'duration_secs': 0.453379} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.591119] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 879.591649] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.591819] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.592386] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.592823] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38b9d6ee-9523-49b4-8711-06cf3bbbe275 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.600822] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 879.600822] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521e7517-d583-4e48-7407-db101a03158b" [ 879.600822] env[62552]: _type = "Task" [ 879.600822] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.608087] env[62552]: DEBUG oslo_vmware.api [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239584, 'name': ReconfigVM_Task, 'duration_secs': 0.190733} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.608862] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267439', 'volume_id': 'f925d139-4314-4651-be8c-f07a21f94eea', 'name': 'volume-f925d139-4314-4651-be8c-f07a21f94eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80478878-ff82-4ed6-a851-8eb2bec01e22', 'attached_at': '', 'detached_at': '', 'volume_id': 'f925d139-4314-4651-be8c-f07a21f94eea', 'serial': 'f925d139-4314-4651-be8c-f07a21f94eea'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 879.614411] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521e7517-d583-4e48-7407-db101a03158b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.686529] env[62552]: DEBUG nova.objects.base [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Object Instance<88d0f502-fc3f-429e-bdf6-a1ebacec2117> lazy-loaded attributes: flavor,info_cache {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 879.721415] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.656s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.722185] env[62552]: DEBUG nova.compute.manager [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 879.726309] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 32.110s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.733145] env[62552]: DEBUG nova.compute.manager [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 879.767547] env[62552]: DEBUG oslo_concurrency.lockutils [req-c229c81b-a6b9-429b-8462-863a676206a2 req-375e52d2-3aa6-4792-8eed-195b9aeb883a service nova] Releasing lock "refresh_cache-526d0f25-3e99-4558-94c0-754ec2a80bad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.767831] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "refresh_cache-526d0f25-3e99-4558-94c0-754ec2a80bad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.767966] env[62552]: DEBUG nova.network.neutron [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.883024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.111802] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521e7517-d583-4e48-7407-db101a03158b, 'name': SearchDatastore_Task, 'duration_secs': 0.026216} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.112388] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.114114] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.114114] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.114114] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.114114] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.114114] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae4d6a9a-de37-41fe-850c-99308f4c846f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.121798] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.122066] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.122772] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b66dd5c-77f5-4695-8a1f-9574a065f518 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.129153] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 880.129153] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52735d97-710f-9ccb-2af3-450eb57b6c7b" [ 880.129153] env[62552]: _type = "Task" [ 880.129153] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.138021] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52735d97-710f-9ccb-2af3-450eb57b6c7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.157433] env[62552]: DEBUG nova.objects.instance [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lazy-loading 'flavor' on Instance uuid 80478878-ff82-4ed6-a851-8eb2bec01e22 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.227658] env[62552]: DEBUG nova.compute.utils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.233228] env[62552]: DEBUG nova.compute.manager [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 880.233432] env[62552]: DEBUG nova.network.neutron [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.239017] env[62552]: INFO nova.compute.claims [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.248612] env[62552]: DEBUG nova.compute.manager [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 880.284017] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.338832] env[62552]: DEBUG nova.network.neutron [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.356144] env[62552]: DEBUG nova.network.neutron [-] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.389400] env[62552]: DEBUG nova.policy [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21a3024f41634f18a55c4da7a2cc7548', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d97468cd7c140ed8ec157434f943d37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.641383] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52735d97-710f-9ccb-2af3-450eb57b6c7b, 'name': SearchDatastore_Task, 'duration_secs': 0.013718} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.642251] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7819b53-29b5-4ee5-99d4-88552d0fbe09 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.647370] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 880.647370] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52df701b-bb0e-bd37-60b2-48c80938f084" [ 880.647370] env[62552]: _type = "Task" [ 880.647370] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.655637] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52df701b-bb0e-bd37-60b2-48c80938f084, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.750015] env[62552]: INFO nova.compute.resource_tracker [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating resource usage from migration ac053045-ef61-4508-bfae-a9c3df4952ed [ 880.840063] env[62552]: DEBUG nova.network.neutron [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Updating instance_info_cache with network_info: [{"id": "e28a5083-5362-482d-a79c-f2103aec3b9f", "address": "fa:16:3e:9d:b5:7a", "network": {"id": "6ad33460-d660-442d-a742-61dcc66532b5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1461793279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f13a11803d1b44178f053cc0a4affedd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape28a5083-53", "ovs_interfaceid": "e28a5083-5362-482d-a79c-f2103aec3b9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.859492] env[62552]: INFO nova.compute.manager [-] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Took 1.42 seconds to deallocate network for instance. [ 880.898163] env[62552]: DEBUG nova.network.neutron [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.952986] env[62552]: DEBUG nova.compute.manager [req-0d421b97-50d6-44fb-abbd-7a395c4fe1a5 req-584ff4c0-86a0-4357-bd5a-ffad04228a4f service nova] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Received event network-vif-deleted-e421f4d6-d892-4c96-8c11-32f1b7553780 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 881.140914] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67062752-c127-41dd-8592-22373b3a5acf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.158350] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45bafe03-fa78-46f4-b33a-0ada8d7254eb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.165678] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72b5c74c-6aa5-4c8b-ad85-efd161e5ed01 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.306s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.166967] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52df701b-bb0e-bd37-60b2-48c80938f084, 'name': SearchDatastore_Task, 'duration_secs': 0.014055} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.192998] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.193119] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 5cee4242-49ae-4cb7-a208-e2982f52fbad/5cee4242-49ae-4cb7-a208-e2982f52fbad.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 881.193666] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17238b21-cc2c-4067-8d71-d137ec2e2c11 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.196340] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df43115-7b7d-4d5a-b838-959c27932667 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.205755] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e616bf9f-2b3d-457b-9671-d4bc33b8c384 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.210814] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 881.210814] env[62552]: value = "task-1239585" [ 881.210814] env[62552]: _type = "Task" [ 881.210814] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.222872] env[62552]: DEBUG nova.compute.provider_tree [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.230282] env[62552]: DEBUG nova.network.neutron [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Successfully created port: 56e2a627-1966-4c66-b733-d9e3cf3435ef {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.232370] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239585, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.244742] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "fea83dde-3181-49dd-a000-5e5cffc8de95" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.248041] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fea83dde-3181-49dd-a000-5e5cffc8de95" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.248041] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "fea83dde-3181-49dd-a000-5e5cffc8de95-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.248041] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fea83dde-3181-49dd-a000-5e5cffc8de95-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.248041] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fea83dde-3181-49dd-a000-5e5cffc8de95-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.248041] env[62552]: INFO nova.compute.manager [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Terminating instance [ 881.264323] env[62552]: DEBUG nova.compute.manager [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 881.289292] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.289566] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.289728] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.289911] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.290074] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.290230] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.290449] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.290612] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.290782] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.290974] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.291182] env[62552]: DEBUG nova.virt.hardware [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.292061] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9824264a-4445-4503-80b8-597155511c7a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.302777] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056dd711-f3df-4468-af06-67bd97f7b450 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.343976] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "refresh_cache-526d0f25-3e99-4558-94c0-754ec2a80bad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.344344] env[62552]: DEBUG nova.compute.manager [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Instance network_info: |[{"id": "e28a5083-5362-482d-a79c-f2103aec3b9f", "address": "fa:16:3e:9d:b5:7a", "network": {"id": "6ad33460-d660-442d-a742-61dcc66532b5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1461793279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f13a11803d1b44178f053cc0a4affedd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape28a5083-53", "ovs_interfaceid": "e28a5083-5362-482d-a79c-f2103aec3b9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 881.344785] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:b5:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e28a5083-5362-482d-a79c-f2103aec3b9f', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.353471] env[62552]: DEBUG oslo.service.loopingcall [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.354380] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.354380] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d5cfb25-081d-4f52-9179-194505fd8d15 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.370404] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.376270] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.376270] env[62552]: value = "task-1239586" [ 881.376270] env[62552]: _type = "Task" [ 881.376270] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.389352] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239586, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.401093] env[62552]: DEBUG oslo_concurrency.lockutils [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.720420] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239585, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.726572] env[62552]: DEBUG nova.scheduler.client.report [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 881.751393] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "refresh_cache-fea83dde-3181-49dd-a000-5e5cffc8de95" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.751499] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired lock "refresh_cache-fea83dde-3181-49dd-a000-5e5cffc8de95" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.751617] env[62552]: DEBUG nova.network.neutron [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.888630] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239586, 'name': CreateVM_Task, 'duration_secs': 0.397094} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.888851] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.889751] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.889904] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "[datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.890325] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.890627] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7333ca46-b848-433a-8973-8d834a0e854e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.895406] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 881.895406] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5229ca0e-f73e-2c02-8c54-38479b60d07c" [ 881.895406] env[62552]: _type = "Task" [ 881.895406] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.905141] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5229ca0e-f73e-2c02-8c54-38479b60d07c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.220750] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239585, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.603994} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.221043] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 5cee4242-49ae-4cb7-a208-e2982f52fbad/5cee4242-49ae-4cb7-a208-e2982f52fbad.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 882.221279] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.221568] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7da1bba5-7d4a-44a1-a53e-9967202b4437 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.227668] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 882.227668] env[62552]: value = "task-1239587" [ 882.227668] env[62552]: _type = "Task" [ 882.227668] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.235514] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.236861] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.510s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.236861] env[62552]: INFO nova.compute.manager [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Migrating [ 882.236861] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.237027] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.238409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.135s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.238673] env[62552]: DEBUG nova.objects.instance [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lazy-loading 'resources' on Instance uuid 474dfa9e-d003-478f-b48f-09e0c0452ffd {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.273644] env[62552]: DEBUG nova.network.neutron [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.383476] env[62552]: DEBUG nova.network.neutron [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.406091] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.406685] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "[datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.406806] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Processing image 35ac675b-d35f-462e-aaa2-a2061f578780 {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.407045] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780/35ac675b-d35f-462e-aaa2-a2061f578780.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.407250] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "[datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780/35ac675b-d35f-462e-aaa2-a2061f578780.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.407449] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.407720] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ed6264b-05bc-4a4a-955a-9cc64f8c0366 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.409410] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40e1004e-1db0-4bd5-82d4-356a44a08957 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.418913] env[62552]: DEBUG oslo_vmware.api [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 882.418913] env[62552]: value = "task-1239588" [ 882.418913] env[62552]: _type = "Task" [ 882.418913] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.422945] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.423147] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.424121] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4709ae2-b2ae-4ee5-8649-969744f9692e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.429874] env[62552]: DEBUG oslo_vmware.api [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239588, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.432788] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 882.432788] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520f0e39-f903-18d2-acf3-3e9b2d602735" [ 882.432788] env[62552]: _type = "Task" [ 882.432788] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.442159] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520f0e39-f903-18d2-acf3-3e9b2d602735, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.738536] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078526} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.738877] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.739850] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2df958e-cd7c-431f-8fae-96f70965816c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.745406] env[62552]: INFO nova.compute.rpcapi [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 882.745965] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.791603] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 5cee4242-49ae-4cb7-a208-e2982f52fbad/5cee4242-49ae-4cb7-a208-e2982f52fbad.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.795673] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-176f65d0-37b3-4b50-85f6-7db2af0b88fc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.816096] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 882.816096] env[62552]: value = "task-1239589" [ 882.816096] env[62552]: _type = "Task" [ 882.816096] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.824832] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239589, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.887376] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Releasing lock "refresh_cache-fea83dde-3181-49dd-a000-5e5cffc8de95" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.887823] env[62552]: DEBUG nova.compute.manager [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 882.888033] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.888923] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab9eadd-e440-417a-a1c0-64b83b3d45a0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.901310] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.901513] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79db60bd-72be-4419-b2f6-aeb099cf0dc7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.908223] env[62552]: DEBUG oslo_vmware.api [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 882.908223] env[62552]: value = "task-1239590" [ 882.908223] env[62552]: _type = "Task" [ 882.908223] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.922443] env[62552]: DEBUG oslo_vmware.api [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239590, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.932641] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.932965] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.940579] env[62552]: DEBUG oslo_vmware.api [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239588, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.950522] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Preparing fetch location {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 882.950857] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Fetch image to [datastore2] OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724/OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724.vmdk {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 882.951100] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Downloading stream optimized image 35ac675b-d35f-462e-aaa2-a2061f578780 to [datastore2] OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724/OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724.vmdk on the data store datastore2 as vApp {{(pid=62552) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 882.951323] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Downloading image file data 35ac675b-d35f-462e-aaa2-a2061f578780 to the ESX as VM named 'OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724' {{(pid=62552) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 883.038310] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 883.038310] env[62552]: value = "resgroup-9" [ 883.038310] env[62552]: _type = "ResourcePool" [ 883.038310] env[62552]: }. {{(pid=62552) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 883.038674] env[62552]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-5b3919d8-052d-4e9d-8858-3a19cdf1b796 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.065446] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lease: (returnval){ [ 883.065446] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52180432-2966-663b-30e8-5b6f410a723b" [ 883.065446] env[62552]: _type = "HttpNfcLease" [ 883.065446] env[62552]: } obtained for vApp import into resource pool (val){ [ 883.065446] env[62552]: value = "resgroup-9" [ 883.065446] env[62552]: _type = "ResourcePool" [ 883.065446] env[62552]: }. {{(pid=62552) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 883.065777] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the lease: (returnval){ [ 883.065777] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52180432-2966-663b-30e8-5b6f410a723b" [ 883.065777] env[62552]: _type = "HttpNfcLease" [ 883.065777] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 883.073077] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.073077] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52180432-2966-663b-30e8-5b6f410a723b" [ 883.073077] env[62552]: _type = "HttpNfcLease" [ 883.073077] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 883.216299] env[62552]: DEBUG nova.compute.manager [req-cbcd0948-c9a6-4ce3-9273-8ad7db547b8b req-e175e97b-cf51-4387-b1bb-9a476b69393f service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Received event network-vif-plugged-56e2a627-1966-4c66-b733-d9e3cf3435ef {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 883.216299] env[62552]: DEBUG oslo_concurrency.lockutils [req-cbcd0948-c9a6-4ce3-9273-8ad7db547b8b req-e175e97b-cf51-4387-b1bb-9a476b69393f service nova] Acquiring lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.216299] env[62552]: DEBUG oslo_concurrency.lockutils [req-cbcd0948-c9a6-4ce3-9273-8ad7db547b8b req-e175e97b-cf51-4387-b1bb-9a476b69393f service nova] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.216299] env[62552]: DEBUG oslo_concurrency.lockutils [req-cbcd0948-c9a6-4ce3-9273-8ad7db547b8b req-e175e97b-cf51-4387-b1bb-9a476b69393f service nova] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.216299] env[62552]: DEBUG nova.compute.manager [req-cbcd0948-c9a6-4ce3-9273-8ad7db547b8b req-e175e97b-cf51-4387-b1bb-9a476b69393f service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] No waiting events found dispatching network-vif-plugged-56e2a627-1966-4c66-b733-d9e3cf3435ef {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 883.216299] env[62552]: WARNING nova.compute.manager [req-cbcd0948-c9a6-4ce3-9273-8ad7db547b8b req-e175e97b-cf51-4387-b1bb-9a476b69393f service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Received unexpected event network-vif-plugged-56e2a627-1966-4c66-b733-d9e3cf3435ef for instance with vm_state building and task_state spawning. [ 883.268923] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.269201] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.269518] env[62552]: DEBUG nova.network.neutron [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.297726] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42503742-ea48-4d32-9795-3f228e001810 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.306534] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98bf31f1-2b4f-44e8-9970-1d689e0f57e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.340222] env[62552]: DEBUG nova.network.neutron [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Successfully updated port: 56e2a627-1966-4c66-b733-d9e3cf3435ef {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.344982] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952e4e71-6a53-4642-aab8-4fc3e9b6ff08 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.355947] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239589, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.357711] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9df2d6c-e21d-46d2-b18a-d212f31a2148 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.371982] env[62552]: DEBUG nova.compute.provider_tree [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.421797] env[62552]: DEBUG oslo_vmware.api [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239590, 'name': PowerOffVM_Task, 'duration_secs': 0.155551} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.426689] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 883.427018] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 883.427511] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c119368f-7874-4e1d-a237-eaec531f37fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.439561] env[62552]: DEBUG oslo_vmware.api [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239588, 'name': PowerOnVM_Task, 'duration_secs': 0.924237} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.439561] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.439561] env[62552]: DEBUG nova.compute.manager [None req-069a75ab-0ba5-4389-ad7f-467d3bc5bdf1 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 883.440855] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325178c9-7dd9-443b-b75a-7a9c93125a29 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.451267] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 883.451528] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 883.451750] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Deleting the datastore file [datastore2] fea83dde-3181-49dd-a000-5e5cffc8de95 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.452259] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-812f079b-be91-4e60-ad34-6110fec0b953 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.458607] env[62552]: DEBUG oslo_vmware.api [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 883.458607] env[62552]: value = "task-1239593" [ 883.458607] env[62552]: _type = "Task" [ 883.458607] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.467514] env[62552]: DEBUG oslo_vmware.api [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239593, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.574776] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.574776] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52180432-2966-663b-30e8-5b6f410a723b" [ 883.574776] env[62552]: _type = "HttpNfcLease" [ 883.574776] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 883.575101] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 883.575101] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52180432-2966-663b-30e8-5b6f410a723b" [ 883.575101] env[62552]: _type = "HttpNfcLease" [ 883.575101] env[62552]: }. {{(pid=62552) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 883.575862] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52104c9a-97f4-4087-bb55-b06702e563f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.583754] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b8323-decf-f9d4-54f3-4e54832407fe/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 883.584098] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b8323-decf-f9d4-54f3-4e54832407fe/disk-0.vmdk. {{(pid=62552) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 883.650175] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d9de6545-335c-4a96-8bb2-a6fc9e13a86e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.852567] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.852880] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.852916] env[62552]: DEBUG nova.network.neutron [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.854524] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239589, 'name': ReconfigVM_Task, 'duration_secs': 0.538918} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.857538] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 5cee4242-49ae-4cb7-a208-e2982f52fbad/5cee4242-49ae-4cb7-a208-e2982f52fbad.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.858554] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07f940d0-8a99-434d-9ad5-88b4672c34e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.873399] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 883.873399] env[62552]: value = "task-1239594" [ 883.873399] env[62552]: _type = "Task" [ 883.873399] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.874727] env[62552]: DEBUG nova.scheduler.client.report [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 883.891345] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239594, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.971298] env[62552]: DEBUG oslo_vmware.api [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239593, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249754} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.972115] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.972756] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 883.973022] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.973249] env[62552]: INFO nova.compute.manager [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Took 1.09 seconds to destroy the instance on the hypervisor. [ 883.973482] env[62552]: DEBUG oslo.service.loopingcall [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.976571] env[62552]: DEBUG nova.compute.manager [-] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 883.976703] env[62552]: DEBUG nova.network.neutron [-] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 884.261390] env[62552]: DEBUG nova.network.neutron [-] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.361947] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Completed reading data from the image iterator. {{(pid=62552) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 884.362271] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b8323-decf-f9d4-54f3-4e54832407fe/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 884.363462] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a51d7e-76d7-4cab-bd56-c4fb283c441a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.374882] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b8323-decf-f9d4-54f3-4e54832407fe/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 884.375147] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b8323-decf-f9d4-54f3-4e54832407fe/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 884.380806] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9b5c1b0e-11fb-485c-97ce-7ca8e5f6108d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.386019] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239594, 'name': Rename_Task, 'duration_secs': 0.205625} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.386114] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.148s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.388177] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.389197] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.833s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.390008] env[62552]: DEBUG nova.objects.instance [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lazy-loading 'resources' on Instance uuid c52efdeb-9efc-4ba1-b9af-4dd0a925517f {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.391243] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f970148d-0465-4653-914f-1e23c44d25af {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.403706] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 884.403706] env[62552]: value = "task-1239595" [ 884.403706] env[62552]: _type = "Task" [ 884.403706] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.413748] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239595, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.421858] env[62552]: INFO nova.scheduler.client.report [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Deleted allocations for instance 474dfa9e-d003-478f-b48f-09e0c0452ffd [ 884.447151] env[62552]: DEBUG nova.network.neutron [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.554240] env[62552]: DEBUG oslo_vmware.rw_handles [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526b8323-decf-f9d4-54f3-4e54832407fe/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 884.554517] env[62552]: INFO nova.virt.vmwareapi.images [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Downloaded image file data 35ac675b-d35f-462e-aaa2-a2061f578780 [ 884.555851] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130abc5e-b00d-46bb-bff4-3a72f92618fb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.579753] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1b3c49a-f7fe-4c01-a7ae-92eedd50d7ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.607660] env[62552]: INFO nova.virt.vmwareapi.images [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] The imported VM was unregistered [ 884.610795] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Caching image {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 884.611169] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Creating directory with path [datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780 {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.611500] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-542c4358-be03-40a2-ae6e-e871bd0135ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.626744] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Created directory with path [datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780 {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.627018] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724/OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724.vmdk to [datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780/35ac675b-d35f-462e-aaa2-a2061f578780.vmdk. {{(pid=62552) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 884.627316] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f41b4105-28d3-4fba-b92c-6692a97027b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.636058] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 884.636058] env[62552]: value = "task-1239597" [ 884.636058] env[62552]: _type = "Task" [ 884.636058] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.646606] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239597, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.764175] env[62552]: DEBUG nova.network.neutron [-] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.792354] env[62552]: DEBUG nova.network.neutron [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance_info_cache with network_info: [{"id": "f749d931-7a86-41ca-a551-4f8d2013603a", "address": "fa:16:3e:00:02:b8", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf749d931-7a", "ovs_interfaceid": "f749d931-7a86-41ca-a551-4f8d2013603a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.920570] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239595, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.932888] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d350c2e8-88af-42ab-88ea-3f36d00655a1 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "474dfa9e-d003-478f-b48f-09e0c0452ffd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.263s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.968781] env[62552]: DEBUG nova.network.neutron [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updating instance_info_cache with network_info: [{"id": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "address": "fa:16:3e:fc:98:ec", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56e2a627-19", "ovs_interfaceid": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.150395] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239597, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.268367] env[62552]: INFO nova.compute.manager [-] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Took 1.29 seconds to deallocate network for instance. [ 885.294892] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.390587] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc094636-74fc-4030-9f58-c5d6116ac4de {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.399266] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72571a76-2537-42bd-bffe-50f5ab59949d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.439174] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e4ce1b-4a59-4d82-9a02-15ff4fbceb61 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.446137] env[62552]: DEBUG oslo_vmware.api [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239595, 'name': PowerOnVM_Task, 'duration_secs': 0.78733} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.446839] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.447170] env[62552]: INFO nova.compute.manager [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Took 11.03 seconds to spawn the instance on the hypervisor. [ 885.447436] env[62552]: DEBUG nova.compute.manager [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 885.448314] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880c7cdb-e908-4d64-915c-e619f9d71ea6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.455835] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f97193-8eaa-42d2-87ba-e295b5f4d808 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.478183] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Releasing lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.478183] env[62552]: DEBUG nova.compute.manager [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Instance network_info: |[{"id": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "address": "fa:16:3e:fc:98:ec", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56e2a627-19", "ovs_interfaceid": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 885.479805] env[62552]: DEBUG nova.compute.provider_tree [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.479879] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:98:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0636c3f6-fcb7-4954-ab07-c5cd0dee37b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56e2a627-1966-4c66-b733-d9e3cf3435ef', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.489993] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Creating folder: Project (8d97468cd7c140ed8ec157434f943d37). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 885.490711] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4cfbca83-45b2-4ec0-b3d0-bb631d402965 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.503432] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Created folder: Project (8d97468cd7c140ed8ec157434f943d37) in parent group-v267339. [ 885.503432] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Creating folder: Instances. Parent ref: group-v267451. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 885.503432] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5ab7f18-d62b-4fbc-ab18-1d59adbf89da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.509243] env[62552]: DEBUG nova.compute.manager [req-b2af1e65-d472-4c75-869a-686646d32fcd req-b09ffb8d-046a-45fc-a234-b9941040a766 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Received event network-changed-56e2a627-1966-4c66-b733-d9e3cf3435ef {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 885.509515] env[62552]: DEBUG nova.compute.manager [req-b2af1e65-d472-4c75-869a-686646d32fcd req-b09ffb8d-046a-45fc-a234-b9941040a766 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Refreshing instance network info cache due to event network-changed-56e2a627-1966-4c66-b733-d9e3cf3435ef. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 885.509791] env[62552]: DEBUG oslo_concurrency.lockutils [req-b2af1e65-d472-4c75-869a-686646d32fcd req-b09ffb8d-046a-45fc-a234-b9941040a766 service nova] Acquiring lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.510400] env[62552]: DEBUG oslo_concurrency.lockutils [req-b2af1e65-d472-4c75-869a-686646d32fcd req-b09ffb8d-046a-45fc-a234-b9941040a766 service nova] Acquired lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.510400] env[62552]: DEBUG nova.network.neutron [req-b2af1e65-d472-4c75-869a-686646d32fcd req-b09ffb8d-046a-45fc-a234-b9941040a766 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Refreshing network info cache for port 56e2a627-1966-4c66-b733-d9e3cf3435ef {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.513727] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Created folder: Instances in parent group-v267451. [ 885.514050] env[62552]: DEBUG oslo.service.loopingcall [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.514683] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 885.515184] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f2ac944-0da7-46a9-adbb-7cb1d61a12b3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.535513] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.535513] env[62552]: value = "task-1239600" [ 885.535513] env[62552]: _type = "Task" [ 885.535513] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.544027] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239600, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.646759] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239597, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.777410] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.976166] env[62552]: INFO nova.compute.manager [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Took 45.58 seconds to build instance. [ 885.992120] env[62552]: DEBUG nova.scheduler.client.report [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 886.046162] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239600, 'name': CreateVM_Task, 'duration_secs': 0.366745} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.046993] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 886.047769] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.047962] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.049209] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 886.051464] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89b96e44-9589-40e9-9cae-bf9a3d4334b4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.057988] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 886.057988] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b0bd21-4245-cc8a-9abe-78adc53cc708" [ 886.057988] env[62552]: _type = "Task" [ 886.057988] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.070719] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b0bd21-4245-cc8a-9abe-78adc53cc708, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.147401] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239597, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.458026] env[62552]: DEBUG nova.network.neutron [req-b2af1e65-d472-4c75-869a-686646d32fcd req-b09ffb8d-046a-45fc-a234-b9941040a766 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updated VIF entry in instance network info cache for port 56e2a627-1966-4c66-b733-d9e3cf3435ef. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.461856] env[62552]: DEBUG nova.network.neutron [req-b2af1e65-d472-4c75-869a-686646d32fcd req-b09ffb8d-046a-45fc-a234-b9941040a766 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updating instance_info_cache with network_info: [{"id": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "address": "fa:16:3e:fc:98:ec", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56e2a627-19", "ovs_interfaceid": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.478434] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f9b49f1e-523a-44d4-b24c-ae1c9d72a935 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "5cee4242-49ae-4cb7-a208-e2982f52fbad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.192s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.497569] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.109s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.500072] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.307s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.501650] env[62552]: INFO nova.compute.claims [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 886.525558] env[62552]: INFO nova.scheduler.client.report [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Deleted allocations for instance c52efdeb-9efc-4ba1-b9af-4dd0a925517f [ 886.537158] env[62552]: INFO nova.compute.manager [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Rescuing [ 886.537158] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.537158] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.537158] env[62552]: DEBUG nova.network.neutron [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.568764] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b0bd21-4245-cc8a-9abe-78adc53cc708, 'name': SearchDatastore_Task, 'duration_secs': 0.029865} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.569450] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.569835] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.570212] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.570486] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.572029] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.572029] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c23ea4ca-1580-4677-be03-4f8c5c1fafe2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.582621] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.583128] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 886.583981] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5da3a40e-22ca-4eab-b348-f90530d154a2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.591036] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 886.591036] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5295d189-b309-ccb0-6b84-c006e446bc59" [ 886.591036] env[62552]: _type = "Task" [ 886.591036] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.600151] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5295d189-b309-ccb0-6b84-c006e446bc59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.648236] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239597, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.810102] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e82e64-bd64-4fdb-9e0b-3c6b4523a80b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.829159] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance '22357d4e-9771-477c-9fc3-fe3d76f6e902' progress to 0 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 886.961661] env[62552]: DEBUG oslo_concurrency.lockutils [req-b2af1e65-d472-4c75-869a-686646d32fcd req-b09ffb8d-046a-45fc-a234-b9941040a766 service nova] Releasing lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.983241] env[62552]: DEBUG nova.compute.manager [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 887.037533] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94d4fb8f-7920-46f7-888c-1293061ec5b2 tempest-MultipleCreateTestJSON-181606411 tempest-MultipleCreateTestJSON-181606411-project-member] Lock "c52efdeb-9efc-4ba1-b9af-4dd0a925517f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.289s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.103453] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5295d189-b309-ccb0-6b84-c006e446bc59, 'name': SearchDatastore_Task, 'duration_secs': 0.022043} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.104422] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1ef058c-0dfb-477a-ae12-c5589edda24c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.110494] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 887.110494] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5229a0d0-ca72-c8e4-5148-e47104a88dec" [ 887.110494] env[62552]: _type = "Task" [ 887.110494] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.129511] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5229a0d0-ca72-c8e4-5148-e47104a88dec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.148422] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239597, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.335748] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.336083] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ccea8a92-64d9-4675-b304-47daeb53590b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.343873] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 887.343873] env[62552]: value = "task-1239601" [ 887.343873] env[62552]: _type = "Task" [ 887.343873] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.357114] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239601, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.509894] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.549834] env[62552]: DEBUG nova.network.neutron [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Updating instance_info_cache with network_info: [{"id": "cf27f46e-a6d0-4547-9eea-2a88fb1f21dc", "address": "fa:16:3e:04:e5:1a", "network": {"id": "925fdab2-f263-46f3-ab17-2f51407bae3e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1178354148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1716b64900f44a8fb2282801905efd67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf27f46e-a6", "ovs_interfaceid": "cf27f46e-a6d0-4547-9eea-2a88fb1f21dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.624584] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5229a0d0-ca72-c8e4-5148-e47104a88dec, 'name': SearchDatastore_Task, 'duration_secs': 0.02249} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.625280] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.625608] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0/a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 887.626225] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-454891b6-b4f1-4998-8e4c-b1b6c6a3b61f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.634119] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 887.634119] env[62552]: value = "task-1239602" [ 887.634119] env[62552]: _type = "Task" [ 887.634119] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.642695] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239602, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.653986] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239597, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.873565} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.654690] env[62552]: INFO nova.virt.vmwareapi.ds_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724/OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724.vmdk to [datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780/35ac675b-d35f-462e-aaa2-a2061f578780.vmdk. [ 887.654951] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Cleaning up location [datastore2] OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724 {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 887.655216] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_4671ad85-4cfe-433b-9b46-8e945cade724 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.655505] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1142aa50-bbf3-4887-bc2c-57f9b621dbb1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.662273] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 887.662273] env[62552]: value = "task-1239603" [ 887.662273] env[62552]: _type = "Task" [ 887.662273] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.675761] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.858782] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239601, 'name': PowerOffVM_Task, 'duration_secs': 0.489469} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.859080] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.859300] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance '22357d4e-9771-477c-9fc3-fe3d76f6e902' progress to 17 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 887.922673] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8208b068-f457-45ed-bc77-35dbb35a743d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.931537] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781afc47-dea9-4f78-9ec7-4a28e7c0beca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.968910] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fdec1c3-70e2-4bff-a1bb-2ae2ae042006 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.976836] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c93006-8b9d-4197-90d6-e5f7f776036f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.992626] env[62552]: DEBUG nova.compute.provider_tree [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.055856] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.148553] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239602, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.179734] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.056326} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.179734] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.179734] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "[datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780/35ac675b-d35f-462e-aaa2-a2061f578780.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.180171] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780/35ac675b-d35f-462e-aaa2-a2061f578780.vmdk to [datastore2] 526d0f25-3e99-4558-94c0-754ec2a80bad/526d0f25-3e99-4558-94c0-754ec2a80bad.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.180538] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5fef23db-1208-4e50-b7df-e64c3f4cf94b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.191157] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 888.191157] env[62552]: value = "task-1239604" [ 888.191157] env[62552]: _type = "Task" [ 888.191157] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.199641] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.365724] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 888.366013] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 888.366188] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.366377] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 888.366529] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.366681] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 888.366891] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 888.367238] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 888.367461] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 888.367636] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 888.367816] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.373153] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fab37546-e9a4-4ca0-a3b3-54c4d9d0fd5c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.390083] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 888.390083] env[62552]: value = "task-1239605" [ 888.390083] env[62552]: _type = "Task" [ 888.390083] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.398782] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.496289] env[62552]: DEBUG nova.scheduler.client.report [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 888.648729] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239602, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582696} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.649049] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0/a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 888.649292] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 888.649573] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-658fd1b7-a851-4a64-8f3c-eae27457889c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.656347] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 888.656347] env[62552]: value = "task-1239606" [ 888.656347] env[62552]: _type = "Task" [ 888.656347] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.664952] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239606, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.705740] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239604, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.905540] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239605, 'name': ReconfigVM_Task, 'duration_secs': 0.377123} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.906310] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance '22357d4e-9771-477c-9fc3-fe3d76f6e902' progress to 33 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 889.002267] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.003073] env[62552]: DEBUG nova.compute.manager [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 889.009572] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.679s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.009877] env[62552]: DEBUG nova.objects.instance [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lazy-loading 'resources' on Instance uuid 51bbcc4e-8251-4b38-9d36-8aea8fc7705d {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.093503] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.094464] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f83908ad-1b26-49f1-a05d-5953524cfb43 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.104798] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 889.104798] env[62552]: value = "task-1239607" [ 889.104798] env[62552]: _type = "Task" [ 889.104798] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.114879] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.168600] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239606, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125466} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.168600] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.169292] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887e5f68-787c-4cc9-a2c8-6b23ad948edd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.193878] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0/a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.194325] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b791d26-87e6-4e11-b92d-9ea5adca2504 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.217890] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 889.217890] env[62552]: value = "task-1239608" [ 889.217890] env[62552]: _type = "Task" [ 889.217890] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.221815] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239604, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.233043] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239608, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.414714] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.414890] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.415014] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.415218] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.415373] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.415527] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.415734] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.415897] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.416078] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.416251] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.416423] env[62552]: DEBUG nova.virt.hardware [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.422046] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Reconfiguring VM instance instance-0000003f to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 889.422460] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-576787ed-d09d-4307-84d7-6d9ec0ccc275 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.442450] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 889.442450] env[62552]: value = "task-1239609" [ 889.442450] env[62552]: _type = "Task" [ 889.442450] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.451850] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239609, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.515935] env[62552]: DEBUG nova.compute.utils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 889.518631] env[62552]: DEBUG nova.compute.manager [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 889.520048] env[62552]: DEBUG nova.network.neutron [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 889.619881] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.659920] env[62552]: DEBUG nova.policy [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80977f84033b4e45a2294b9603f9f358', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f2003df754340ab8e4cf9ca5a21b9a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 889.706648] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239604, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.737235] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239608, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.960082] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.981899] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c0e199-7628-43fd-a764-e3cbbe9dce46 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.991486] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e77a133-1c0d-46f9-9342-b87cf46e7dba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.027998] env[62552]: DEBUG nova.compute.manager [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 890.031843] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c043828c-10de-4288-92ed-73268d312b5d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.042457] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0440b3e-5757-42e2-9acf-9a3b4e80082c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.063654] env[62552]: DEBUG nova.compute.provider_tree [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.120669] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239607, 'name': PowerOffVM_Task, 'duration_secs': 1.016353} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.121200] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.122023] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c281d6f-8a4a-48d8-a025-e1f390850775 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.145147] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a298db-0eb1-44eb-aff7-dcf36e4440ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.183817] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.183891] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78da5a0d-04cd-483f-8047-c6cd31c6407a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.194976] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 890.194976] env[62552]: value = "task-1239610" [ 890.194976] env[62552]: _type = "Task" [ 890.194976] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.215911] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239604, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.219749] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 890.219749] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.219749] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.219749] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.219749] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.219943] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db3ae2f2-0183-4d3a-9960-6eca2ad02886 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.231030] env[62552]: DEBUG nova.network.neutron [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Successfully created port: 123341c8-40fb-4f81-9d1c-1f556b6a92b8 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 890.236647] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239608, 'name': ReconfigVM_Task, 'duration_secs': 0.888264} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.238015] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Reconfigured VM instance instance-00000049 to attach disk [datastore1] a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0/a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.238881] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.239043] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.239861] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-456c3fcf-2926-43b9-9206-db30ba9b505b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.242106] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6b0cd1a-4110-4041-837d-67bdac5b6265 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.247514] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 890.247514] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52711512-d955-0cb5-3d43-d0e25fcf49db" [ 890.247514] env[62552]: _type = "Task" [ 890.247514] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.253150] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 890.253150] env[62552]: value = "task-1239611" [ 890.253150] env[62552]: _type = "Task" [ 890.253150] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.260150] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52711512-d955-0cb5-3d43-d0e25fcf49db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.266590] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239611, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.454930] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.567085] env[62552]: DEBUG nova.scheduler.client.report [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 890.709270] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239604, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.762181] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52711512-d955-0cb5-3d43-d0e25fcf49db, 'name': SearchDatastore_Task, 'duration_secs': 0.095277} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.763435] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20a2e9ab-7e9f-48e5-8770-0fbbe961114b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.768595] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239611, 'name': Rename_Task, 'duration_secs': 0.157145} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.769287] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 890.769575] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1c23201-fc11-4df0-be25-f8c6bc379586 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.773386] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 890.773386] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f66d9d-0f27-ea51-a4c0-ccb54c96573d" [ 890.773386] env[62552]: _type = "Task" [ 890.773386] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.777503] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 890.777503] env[62552]: value = "task-1239612" [ 890.777503] env[62552]: _type = "Task" [ 890.777503] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.784412] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f66d9d-0f27-ea51-a4c0-ccb54c96573d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.789088] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239612, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.956494] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239609, 'name': ReconfigVM_Task, 'duration_secs': 1.123301} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.956862] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Reconfigured VM instance instance-0000003f to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 890.957825] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80b2871-ecec-44a1-b861-290ecaa5f0a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.980994] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 22357d4e-9771-477c-9fc3-fe3d76f6e902/22357d4e-9771-477c-9fc3-fe3d76f6e902.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.981436] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f76e880c-4c12-4d60-b0e2-3e61dfe7f589 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.001398] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 891.001398] env[62552]: value = "task-1239613" [ 891.001398] env[62552]: _type = "Task" [ 891.001398] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.009750] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239613, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.040485] env[62552]: DEBUG nova.compute.manager [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 891.074450] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.074781] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.074976] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.075212] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.075404] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.075571] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.075873] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.076081] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.076354] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.076450] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.076741] env[62552]: DEBUG nova.virt.hardware [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.077610] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.068s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.080301] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed59921-b12b-4ede-8a38-753ddfc3adc3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.083474] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.532s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.085118] env[62552]: INFO nova.compute.claims [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.094501] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6388791f-4e1a-48dc-a475-fb16b8334102 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.113377] env[62552]: INFO nova.scheduler.client.report [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Deleted allocations for instance 51bbcc4e-8251-4b38-9d36-8aea8fc7705d [ 891.208348] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239604, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.597277} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.208806] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/35ac675b-d35f-462e-aaa2-a2061f578780/35ac675b-d35f-462e-aaa2-a2061f578780.vmdk to [datastore2] 526d0f25-3e99-4558-94c0-754ec2a80bad/526d0f25-3e99-4558-94c0-754ec2a80bad.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 891.209710] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c1ac04-2a70-4206-a460-0bf65e268093 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.233771] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 526d0f25-3e99-4558-94c0-754ec2a80bad/526d0f25-3e99-4558-94c0-754ec2a80bad.vmdk or device None with type streamOptimized {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.234517] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec3dadf5-1f3d-42e4-a391-23b0c0b58a45 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.254170] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 891.254170] env[62552]: value = "task-1239614" [ 891.254170] env[62552]: _type = "Task" [ 891.254170] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.267515] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239614, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.286188] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f66d9d-0f27-ea51-a4c0-ccb54c96573d, 'name': SearchDatastore_Task, 'duration_secs': 0.013085} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.286971] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.287351] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 5cee4242-49ae-4cb7-a208-e2982f52fbad/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. {{(pid=62552) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 891.287716] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-232c272c-0aba-4f3b-8b8a-6256d84301af {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.293714] env[62552]: DEBUG oslo_vmware.api [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239612, 'name': PowerOnVM_Task, 'duration_secs': 0.479086} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.294372] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 891.294614] env[62552]: INFO nova.compute.manager [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Took 10.03 seconds to spawn the instance on the hypervisor. [ 891.294807] env[62552]: DEBUG nova.compute.manager [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 891.295628] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6b1ff0-bf42-4997-9c47-0abe506ffd4b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.300697] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 891.300697] env[62552]: value = "task-1239615" [ 891.300697] env[62552]: _type = "Task" [ 891.300697] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.314421] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239615, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.510604] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239613, 'name': ReconfigVM_Task, 'duration_secs': 0.480327} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.510905] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 22357d4e-9771-477c-9fc3-fe3d76f6e902/22357d4e-9771-477c-9fc3-fe3d76f6e902.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.511260] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance '22357d4e-9771-477c-9fc3-fe3d76f6e902' progress to 50 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 891.613581] env[62552]: DEBUG oslo_vmware.rw_handles [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527499e9-eb32-4eef-49a1-e96bfcd355eb/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 891.614610] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc10988-f9b9-4a7d-8671-60c0c7b6ab95 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.623243] env[62552]: DEBUG oslo_vmware.rw_handles [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527499e9-eb32-4eef-49a1-e96bfcd355eb/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 891.623567] env[62552]: ERROR oslo_vmware.rw_handles [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527499e9-eb32-4eef-49a1-e96bfcd355eb/disk-0.vmdk due to incomplete transfer. [ 891.624282] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d86e5405-7b4e-455b-b7b6-5445b43bcd16 tempest-AttachInterfacesUnderV243Test-1521376707 tempest-AttachInterfacesUnderV243Test-1521376707-project-member] Lock "51bbcc4e-8251-4b38-9d36-8aea8fc7705d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.483s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.625150] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c1a42239-1066-41da-8f33-d41cb2816965 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.634484] env[62552]: DEBUG oslo_vmware.rw_handles [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527499e9-eb32-4eef-49a1-e96bfcd355eb/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 891.634714] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Uploaded image fd4b6cf6-b251-4731-93a3-ed4f4acab88b to the Glance image server {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 891.637398] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 891.637698] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-da91ae25-1b48-4f9b-b91d-e8a2cf4a8e17 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.644459] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 891.644459] env[62552]: value = "task-1239616" [ 891.644459] env[62552]: _type = "Task" [ 891.644459] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.654734] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239616, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.766812] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239614, 'name': ReconfigVM_Task, 'duration_secs': 0.324498} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.770531] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 526d0f25-3e99-4558-94c0-754ec2a80bad/526d0f25-3e99-4558-94c0-754ec2a80bad.vmdk or device None with type streamOptimized {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.770531] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4459a92-242f-402a-8d74-8d440d9fb960 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.775761] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 891.775761] env[62552]: value = "task-1239617" [ 891.775761] env[62552]: _type = "Task" [ 891.775761] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.784826] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239617, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.822406] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239615, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.825622] env[62552]: INFO nova.compute.manager [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Took 45.83 seconds to build instance. [ 892.020050] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bfbffa-22d1-4fff-ad5f-b857e7be6a3c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.041959] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78b4101-723d-4d20-8948-0d84174e6cb7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.048026] env[62552]: DEBUG nova.compute.manager [req-0cb43d09-8e93-415e-bdc2-6c1d677784b5 req-f69bb2c8-f864-4294-a4cb-496d22461a3c service nova] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Received event network-vif-plugged-123341c8-40fb-4f81-9d1c-1f556b6a92b8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 892.049016] env[62552]: DEBUG oslo_concurrency.lockutils [req-0cb43d09-8e93-415e-bdc2-6c1d677784b5 req-f69bb2c8-f864-4294-a4cb-496d22461a3c service nova] Acquiring lock "11831580-1b58-476a-91ce-a4e55947fd91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.049016] env[62552]: DEBUG oslo_concurrency.lockutils [req-0cb43d09-8e93-415e-bdc2-6c1d677784b5 req-f69bb2c8-f864-4294-a4cb-496d22461a3c service nova] Lock "11831580-1b58-476a-91ce-a4e55947fd91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.049016] env[62552]: DEBUG oslo_concurrency.lockutils [req-0cb43d09-8e93-415e-bdc2-6c1d677784b5 req-f69bb2c8-f864-4294-a4cb-496d22461a3c service nova] Lock "11831580-1b58-476a-91ce-a4e55947fd91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.049263] env[62552]: DEBUG nova.compute.manager [req-0cb43d09-8e93-415e-bdc2-6c1d677784b5 req-f69bb2c8-f864-4294-a4cb-496d22461a3c service nova] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] No waiting events found dispatching network-vif-plugged-123341c8-40fb-4f81-9d1c-1f556b6a92b8 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 892.049649] env[62552]: WARNING nova.compute.manager [req-0cb43d09-8e93-415e-bdc2-6c1d677784b5 req-f69bb2c8-f864-4294-a4cb-496d22461a3c service nova] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Received unexpected event network-vif-plugged-123341c8-40fb-4f81-9d1c-1f556b6a92b8 for instance with vm_state building and task_state spawning. [ 892.068371] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance '22357d4e-9771-477c-9fc3-fe3d76f6e902' progress to 67 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 892.156031] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239616, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.224917] env[62552]: DEBUG nova.network.neutron [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Successfully updated port: 123341c8-40fb-4f81-9d1c-1f556b6a92b8 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.285145] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239617, 'name': Rename_Task, 'duration_secs': 0.16058} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.287906] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 892.288437] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53e9bef8-5f9f-49b5-bd4b-57730eb5d61c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.294810] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 892.294810] env[62552]: value = "task-1239618" [ 892.294810] env[62552]: _type = "Task" [ 892.294810] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.305010] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239618, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.313281] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239615, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581049} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.315825] env[62552]: INFO nova.virt.vmwareapi.ds_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 5cee4242-49ae-4cb7-a208-e2982f52fbad/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. [ 892.317124] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f4632c-1cbd-42ba-92d7-de7335d58dfc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.336537] env[62552]: DEBUG oslo_concurrency.lockutils [None req-286e7bbb-8884-4fd8-bf4e-794a8d90b05c tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.517s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.344445] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 5cee4242-49ae-4cb7-a208-e2982f52fbad/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.347293] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b45be740-8433-4fb9-9ee2-a790c128e884 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.366121] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 892.366121] env[62552]: value = "task-1239619" [ 892.366121] env[62552]: _type = "Task" [ 892.366121] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.375755] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239619, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.504393] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b623c6e4-d9d6-4c60-ab72-8af23c2b74ca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.516886] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9466a8ea-0413-40aa-b70c-69d42ee6b491 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.550239] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35751930-1b8c-456b-8e51-a4a92b5360e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.557318] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b176827-8e63-415a-a066-6ea2a7eb367c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.570780] env[62552]: DEBUG nova.compute.provider_tree [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.655178] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239616, 'name': Destroy_Task, 'duration_secs': 0.792162} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.655474] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Destroyed the VM [ 892.657117] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 892.657117] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6f72ea50-ec30-4853-b778-6b56a2ff8863 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.662881] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 892.662881] env[62552]: value = "task-1239620" [ 892.662881] env[62552]: _type = "Task" [ 892.662881] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.672890] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239620, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.689783] env[62552]: DEBUG nova.network.neutron [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Port f749d931-7a86-41ca-a551-4f8d2013603a binding to destination host cpu-1 is already ACTIVE {{(pid=62552) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 892.727634] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.727865] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.728010] env[62552]: DEBUG nova.network.neutron [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.806884] env[62552]: DEBUG oslo_vmware.api [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239618, 'name': PowerOnVM_Task, 'duration_secs': 0.491581} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.807284] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.807583] env[62552]: INFO nova.compute.manager [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Took 15.77 seconds to spawn the instance on the hypervisor. [ 892.807978] env[62552]: DEBUG nova.compute.manager [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 892.808810] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8619cf43-edf7-4308-a2c8-d9033c0e772b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.849175] env[62552]: DEBUG nova.compute.manager [req-dcd519a6-47c6-43b0-bea9-c27c3093d9a5 req-1b7c738e-f770-46d9-b3d9-0f2e9cbeaf33 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Received event network-changed-56e2a627-1966-4c66-b733-d9e3cf3435ef {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 892.849808] env[62552]: DEBUG nova.compute.manager [req-dcd519a6-47c6-43b0-bea9-c27c3093d9a5 req-1b7c738e-f770-46d9-b3d9-0f2e9cbeaf33 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Refreshing instance network info cache due to event network-changed-56e2a627-1966-4c66-b733-d9e3cf3435ef. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 892.850482] env[62552]: DEBUG oslo_concurrency.lockutils [req-dcd519a6-47c6-43b0-bea9-c27c3093d9a5 req-1b7c738e-f770-46d9-b3d9-0f2e9cbeaf33 service nova] Acquiring lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.850781] env[62552]: DEBUG oslo_concurrency.lockutils [req-dcd519a6-47c6-43b0-bea9-c27c3093d9a5 req-1b7c738e-f770-46d9-b3d9-0f2e9cbeaf33 service nova] Acquired lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.850989] env[62552]: DEBUG nova.network.neutron [req-dcd519a6-47c6-43b0-bea9-c27c3093d9a5 req-1b7c738e-f770-46d9-b3d9-0f2e9cbeaf33 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Refreshing network info cache for port 56e2a627-1966-4c66-b733-d9e3cf3435ef {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 892.877163] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239619, 'name': ReconfigVM_Task, 'duration_secs': 0.338288} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.877570] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 5cee4242-49ae-4cb7-a208-e2982f52fbad/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.878556] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4add454-809a-43f1-b265-aafe7ddcbedc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.909112] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39131d75-82e5-4f27-aa45-e98e2ab9daea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.927124] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 892.927124] env[62552]: value = "task-1239621" [ 892.927124] env[62552]: _type = "Task" [ 892.927124] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.935773] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239621, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.073915] env[62552]: DEBUG nova.scheduler.client.report [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 893.171931] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239620, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.296273] env[62552]: DEBUG nova.network.neutron [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 893.333940] env[62552]: INFO nova.compute.manager [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Took 49.30 seconds to build instance. [ 893.438106] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239621, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.579313] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.582731] env[62552]: DEBUG nova.compute.manager [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 893.583033] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.202s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.584328] env[62552]: INFO nova.compute.claims [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.676036] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239620, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.715317] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "22357d4e-9771-477c-9fc3-fe3d76f6e902-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.720024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.720024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.837774] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8490500-f102-48b7-9fb9-298091f41bb7 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "526d0f25-3e99-4558-94c0-754ec2a80bad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.817s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.850935] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "526d0f25-3e99-4558-94c0-754ec2a80bad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.851643] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "526d0f25-3e99-4558-94c0-754ec2a80bad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.851894] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "526d0f25-3e99-4558-94c0-754ec2a80bad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.852191] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "526d0f25-3e99-4558-94c0-754ec2a80bad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.852415] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "526d0f25-3e99-4558-94c0-754ec2a80bad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.855890] env[62552]: INFO nova.compute.manager [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Terminating instance [ 893.942952] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239621, 'name': ReconfigVM_Task, 'duration_secs': 0.858857} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.947336] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.947679] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f46f5b97-3744-48b7-ba9b-138d3f0ca789 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.952343] env[62552]: DEBUG nova.network.neutron [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Updating instance_info_cache with network_info: [{"id": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "address": "fa:16:3e:ce:cc:01", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap123341c8-40", "ovs_interfaceid": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.959615] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 893.959615] env[62552]: value = "task-1239622" [ 893.959615] env[62552]: _type = "Task" [ 893.959615] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.969928] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239622, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.089403] env[62552]: DEBUG nova.compute.utils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.090912] env[62552]: DEBUG nova.compute.manager [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 894.091670] env[62552]: DEBUG nova.network.neutron [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 894.175558] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239620, 'name': RemoveSnapshot_Task} progress is 56%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.242781] env[62552]: DEBUG nova.policy [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80977f84033b4e45a2294b9603f9f358', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f2003df754340ab8e4cf9ca5a21b9a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.281306] env[62552]: DEBUG nova.network.neutron [req-dcd519a6-47c6-43b0-bea9-c27c3093d9a5 req-1b7c738e-f770-46d9-b3d9-0f2e9cbeaf33 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updated VIF entry in instance network info cache for port 56e2a627-1966-4c66-b733-d9e3cf3435ef. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 894.281724] env[62552]: DEBUG nova.network.neutron [req-dcd519a6-47c6-43b0-bea9-c27c3093d9a5 req-1b7c738e-f770-46d9-b3d9-0f2e9cbeaf33 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updating instance_info_cache with network_info: [{"id": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "address": "fa:16:3e:fc:98:ec", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56e2a627-19", "ovs_interfaceid": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.344758] env[62552]: DEBUG nova.compute.manager [req-622e5bd8-9cb3-44f4-86ab-5f92ee7c075d req-8514044b-0eec-4fd6-affc-e7b1d3c8b6e0 service nova] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Received event network-changed-123341c8-40fb-4f81-9d1c-1f556b6a92b8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 894.345295] env[62552]: DEBUG nova.compute.manager [req-622e5bd8-9cb3-44f4-86ab-5f92ee7c075d req-8514044b-0eec-4fd6-affc-e7b1d3c8b6e0 service nova] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Refreshing instance network info cache due to event network-changed-123341c8-40fb-4f81-9d1c-1f556b6a92b8. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 894.345295] env[62552]: DEBUG oslo_concurrency.lockutils [req-622e5bd8-9cb3-44f4-86ab-5f92ee7c075d req-8514044b-0eec-4fd6-affc-e7b1d3c8b6e0 service nova] Acquiring lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.364698] env[62552]: DEBUG nova.compute.manager [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 894.364929] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.368389] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65853fca-8c7d-460e-9224-5815007fdc2d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.376236] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.376536] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07a3cc03-1520-4ed1-a3f4-357c29700ac2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.383367] env[62552]: DEBUG oslo_vmware.api [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 894.383367] env[62552]: value = "task-1239623" [ 894.383367] env[62552]: _type = "Task" [ 894.383367] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.394784] env[62552]: DEBUG oslo_vmware.api [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239623, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.455060] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.455947] env[62552]: DEBUG nova.compute.manager [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Instance network_info: |[{"id": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "address": "fa:16:3e:ce:cc:01", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap123341c8-40", "ovs_interfaceid": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 894.455947] env[62552]: DEBUG oslo_concurrency.lockutils [req-622e5bd8-9cb3-44f4-86ab-5f92ee7c075d req-8514044b-0eec-4fd6-affc-e7b1d3c8b6e0 service nova] Acquired lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.456135] env[62552]: DEBUG nova.network.neutron [req-622e5bd8-9cb3-44f4-86ab-5f92ee7c075d req-8514044b-0eec-4fd6-affc-e7b1d3c8b6e0 service nova] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Refreshing network info cache for port 123341c8-40fb-4f81-9d1c-1f556b6a92b8 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 894.457396] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:cc:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '838c9497-35dd-415e-96c7-8dc21b0cd4b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '123341c8-40fb-4f81-9d1c-1f556b6a92b8', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.469321] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Creating folder: Project (4f2003df754340ab8e4cf9ca5a21b9a6). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 894.470604] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a60e02d-8798-4879-910a-4d195ed8cf64 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.482473] env[62552]: DEBUG oslo_vmware.api [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239622, 'name': PowerOnVM_Task, 'duration_secs': 0.468122} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.482473] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.485702] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Created folder: Project (4f2003df754340ab8e4cf9ca5a21b9a6) in parent group-v267339. [ 894.485940] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Creating folder: Instances. Parent ref: group-v267454. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 894.487032] env[62552]: DEBUG nova.compute.manager [None req-3e77f334-c6b6-4dc7-adbe-66a29c200cab tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 894.487319] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b7e3aed-d96e-4968-9920-ae26071f5bdd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.489597] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462b3899-ffce-4b84-9763-88cc0e2d9e6e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.499978] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Created folder: Instances in parent group-v267454. [ 894.500295] env[62552]: DEBUG oslo.service.loopingcall [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.500518] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.500754] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-50778244-1dc4-4ad5-94f1-3536a45684ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.520014] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.520014] env[62552]: value = "task-1239626" [ 894.520014] env[62552]: _type = "Task" [ 894.520014] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.529469] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239626, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.600731] env[62552]: DEBUG nova.compute.manager [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 894.675533] env[62552]: DEBUG oslo_vmware.api [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239620, 'name': RemoveSnapshot_Task, 'duration_secs': 1.810475} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.677360] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 894.677645] env[62552]: INFO nova.compute.manager [None req-9e23dbc1-8104-4dfe-807c-60935b4aee62 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Took 20.80 seconds to snapshot the instance on the hypervisor. [ 894.789021] env[62552]: DEBUG oslo_concurrency.lockutils [req-dcd519a6-47c6-43b0-bea9-c27c3093d9a5 req-1b7c738e-f770-46d9-b3d9-0f2e9cbeaf33 service nova] Releasing lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.821434] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.821629] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.821813] env[62552]: DEBUG nova.network.neutron [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.838550] env[62552]: DEBUG nova.network.neutron [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Successfully created port: 24b537e1-3264-41fe-8165-63c833fc5c62 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.902972] env[62552]: DEBUG oslo_vmware.api [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239623, 'name': PowerOffVM_Task, 'duration_secs': 0.18617} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.903334] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 894.903556] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 894.903837] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f37baa1b-6b92-4440-9106-5a8f1c5ac573 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.029110] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.030119] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.030119] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleting the datastore file [datastore2] 526d0f25-3e99-4558-94c0-754ec2a80bad {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.030119] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf15d612-4e1c-4fd5-9cdc-ba97911e9886 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.037144] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239626, 'name': CreateVM_Task, 'duration_secs': 0.367397} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.037664] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.038714] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.038869] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.039752] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.039752] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49ae6ad1-1773-4cec-aaa2-66ca7e9b707a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.045773] env[62552]: DEBUG oslo_vmware.api [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 895.045773] env[62552]: value = "task-1239628" [ 895.045773] env[62552]: _type = "Task" [ 895.045773] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.053150] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 895.053150] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d07ec9-7349-8f67-857f-68ea62bde780" [ 895.053150] env[62552]: _type = "Task" [ 895.053150] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.061521] env[62552]: DEBUG oslo_vmware.api [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.065926] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d07ec9-7349-8f67-857f-68ea62bde780, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.116622] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5897dcdd-67a0-4876-a903-82e9eb9f7634 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.126777] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32231a8d-714a-4295-8785-e82cf1f58a97 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.160935] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e1d321-26a2-4ea7-a26f-27d8a74fb63d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.169373] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a71d43-e880-4d3d-87ba-f8903fa62b47 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.187085] env[62552]: DEBUG nova.compute.provider_tree [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.447259] env[62552]: INFO nova.compute.manager [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Unrescuing [ 895.447564] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.447721] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquired lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.447891] env[62552]: DEBUG nova.network.neutron [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 895.558272] env[62552]: DEBUG oslo_vmware.api [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.565871] env[62552]: DEBUG nova.network.neutron [req-622e5bd8-9cb3-44f4-86ab-5f92ee7c075d req-8514044b-0eec-4fd6-affc-e7b1d3c8b6e0 service nova] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Updated VIF entry in instance network info cache for port 123341c8-40fb-4f81-9d1c-1f556b6a92b8. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 895.565871] env[62552]: DEBUG nova.network.neutron [req-622e5bd8-9cb3-44f4-86ab-5f92ee7c075d req-8514044b-0eec-4fd6-affc-e7b1d3c8b6e0 service nova] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Updating instance_info_cache with network_info: [{"id": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "address": "fa:16:3e:ce:cc:01", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap123341c8-40", "ovs_interfaceid": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.570535] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d07ec9-7349-8f67-857f-68ea62bde780, 'name': SearchDatastore_Task, 'duration_secs': 0.026828} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.571071] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.571326] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.571568] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.571731] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.572417] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.572713] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-078c2bea-298a-477a-8104-341a29e9e9b6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.581885] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.582505] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.582787] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef528672-c6f6-41fd-a4e8-c46941cbc224 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.589196] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 895.589196] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ed821f-cc85-9669-e997-0bc6f4403317" [ 895.589196] env[62552]: _type = "Task" [ 895.589196] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.596881] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ed821f-cc85-9669-e997-0bc6f4403317, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.612596] env[62552]: DEBUG nova.compute.manager [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 895.644250] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.644479] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.644761] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.645077] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.645345] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.645607] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.645934] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.646229] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.648020] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.648020] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.648020] env[62552]: DEBUG nova.virt.hardware [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.648262] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ae73d3-923a-4ddf-8fb6-a76216ab4082 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.658168] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfdb0a6-977d-4d7e-8556-093b9390b919 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.695177] env[62552]: DEBUG nova.scheduler.client.report [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 895.797569] env[62552]: DEBUG nova.network.neutron [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance_info_cache with network_info: [{"id": "f749d931-7a86-41ca-a551-4f8d2013603a", "address": "fa:16:3e:00:02:b8", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf749d931-7a", "ovs_interfaceid": "f749d931-7a86-41ca-a551-4f8d2013603a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.058462] env[62552]: DEBUG oslo_vmware.api [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.778705} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.058731] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.058924] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 896.059118] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 896.059300] env[62552]: INFO nova.compute.manager [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Took 1.69 seconds to destroy the instance on the hypervisor. [ 896.059542] env[62552]: DEBUG oslo.service.loopingcall [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.059744] env[62552]: DEBUG nova.compute.manager [-] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 896.059862] env[62552]: DEBUG nova.network.neutron [-] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 896.071387] env[62552]: DEBUG oslo_concurrency.lockutils [req-622e5bd8-9cb3-44f4-86ab-5f92ee7c075d req-8514044b-0eec-4fd6-affc-e7b1d3c8b6e0 service nova] Releasing lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.101345] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ed821f-cc85-9669-e997-0bc6f4403317, 'name': SearchDatastore_Task, 'duration_secs': 0.034482} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.102196] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bec8bf3c-5bf0-41d4-a6e0-c90b69a18b30 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.109423] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 896.109423] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529ef20a-e52f-487a-b861-366b8b4e3e12" [ 896.109423] env[62552]: _type = "Task" [ 896.109423] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.118788] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529ef20a-e52f-487a-b861-366b8b4e3e12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.197514] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.614s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.198293] env[62552]: DEBUG nova.compute.manager [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 896.203756] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.189s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.206047] env[62552]: INFO nova.compute.claims [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.278178] env[62552]: DEBUG nova.network.neutron [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Updating instance_info_cache with network_info: [{"id": "cf27f46e-a6d0-4547-9eea-2a88fb1f21dc", "address": "fa:16:3e:04:e5:1a", "network": {"id": "925fdab2-f263-46f3-ab17-2f51407bae3e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1178354148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1716b64900f44a8fb2282801905efd67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf27f46e-a6", "ovs_interfaceid": "cf27f46e-a6d0-4547-9eea-2a88fb1f21dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.299361] env[62552]: DEBUG oslo_concurrency.lockutils [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.623731] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529ef20a-e52f-487a-b861-366b8b4e3e12, 'name': SearchDatastore_Task, 'duration_secs': 0.01179} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.624089] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.624414] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 11831580-1b58-476a-91ce-a4e55947fd91/11831580-1b58-476a-91ce-a4e55947fd91.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 896.625230] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12f56d44-71f0-4683-a2b9-f1187f5b3bb5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.633155] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 896.633155] env[62552]: value = "task-1239629" [ 896.633155] env[62552]: _type = "Task" [ 896.633155] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.642009] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239629, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.705234] env[62552]: DEBUG nova.compute.utils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.706836] env[62552]: DEBUG nova.compute.manager [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 896.707614] env[62552]: DEBUG nova.network.neutron [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.748787] env[62552]: DEBUG nova.compute.manager [req-cfb1d1f1-e0df-4f5e-92c7-f4794d58d7d6 req-da15add0-df58-4af0-ae1c-2a8d92063c17 service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Received event network-vif-deleted-e28a5083-5362-482d-a79c-f2103aec3b9f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 896.749131] env[62552]: INFO nova.compute.manager [req-cfb1d1f1-e0df-4f5e-92c7-f4794d58d7d6 req-da15add0-df58-4af0-ae1c-2a8d92063c17 service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Neutron deleted interface e28a5083-5362-482d-a79c-f2103aec3b9f; detaching it from the instance and deleting it from the info cache [ 896.749434] env[62552]: DEBUG nova.network.neutron [req-cfb1d1f1-e0df-4f5e-92c7-f4794d58d7d6 req-da15add0-df58-4af0-ae1c-2a8d92063c17 service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.781703] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Releasing lock "refresh_cache-5cee4242-49ae-4cb7-a208-e2982f52fbad" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.782452] env[62552]: DEBUG nova.objects.instance [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lazy-loading 'flavor' on Instance uuid 5cee4242-49ae-4cb7-a208-e2982f52fbad {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.814476] env[62552]: DEBUG nova.policy [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8792502911d245868e4780eee01b98eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42f9b9d438854dcd870e6b99cf496307', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.847825] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db99cb55-27a5-4028-a637-bfd430b82f82 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.875173] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6157c0f3-b7fa-40b7-ba07-13d41c60da9a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.885530] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance '22357d4e-9771-477c-9fc3-fe3d76f6e902' progress to 83 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 896.943232] env[62552]: DEBUG nova.network.neutron [-] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.143184] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239629, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4877} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.143506] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 11831580-1b58-476a-91ce-a4e55947fd91/11831580-1b58-476a-91ce-a4e55947fd91.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 897.143709] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.143965] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-727c4f99-77d1-432e-bf0c-387ea16048c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.149821] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 897.149821] env[62552]: value = "task-1239630" [ 897.149821] env[62552]: _type = "Task" [ 897.149821] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.159040] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239630, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.192720] env[62552]: DEBUG nova.network.neutron [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Successfully updated port: 24b537e1-3264-41fe-8165-63c833fc5c62 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 897.211721] env[62552]: DEBUG nova.compute.manager [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 897.252989] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e9f4f5c-fb18-4187-a3ba-3a817c8a9f35 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.262529] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d68a1f-aadd-45e4-b7eb-ef30b6988523 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.276325] env[62552]: DEBUG nova.compute.manager [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 897.277613] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae40119f-c120-4f92-b7f8-465cf9e3f9d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.298205] env[62552]: DEBUG nova.compute.manager [req-cfb1d1f1-e0df-4f5e-92c7-f4794d58d7d6 req-da15add0-df58-4af0-ae1c-2a8d92063c17 service nova] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Detach interface failed, port_id=e28a5083-5362-482d-a79c-f2103aec3b9f, reason: Instance 526d0f25-3e99-4558-94c0-754ec2a80bad could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 897.301747] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9aba187-570a-4c86-980d-eb3bc9610080 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.322682] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.326258] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1aeb34fa-9471-4b58-826b-7d6107a0eb51 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.333472] env[62552]: DEBUG oslo_vmware.api [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 897.333472] env[62552]: value = "task-1239631" [ 897.333472] env[62552]: _type = "Task" [ 897.333472] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.341260] env[62552]: DEBUG oslo_vmware.api [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239631, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.392766] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.393139] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db103ca5-4f02-4b20-a431-147dfb928df8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.401879] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 897.401879] env[62552]: value = "task-1239632" [ 897.401879] env[62552]: _type = "Task" [ 897.401879] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.407742] env[62552]: DEBUG nova.network.neutron [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Successfully created port: a9ece5cd-e814-4895-b895-6ac928e22cfb {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.415320] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239632, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.445688] env[62552]: INFO nova.compute.manager [-] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Took 1.39 seconds to deallocate network for instance. [ 897.625740] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ced7907-c160-45e4-9ed1-7a5699a123e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.633127] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07320689-2e56-4d87-aa0e-73c0a3892e17 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.341329] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "refresh_cache-6c8b5367-4fa0-479b-9382-ff261201e3ef" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.341659] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "refresh_cache-6c8b5367-4fa0-479b-9382-ff261201e3ef" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.341659] env[62552]: DEBUG nova.network.neutron [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 898.348025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7115138-b6f2-47b9-a041-913a90588cc2 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.348025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7115138-b6f2-47b9-a041-913a90588cc2 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.348025] env[62552]: DEBUG nova.objects.instance [None req-e7115138-b6f2-47b9-a041-913a90588cc2 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'flavor' on Instance uuid e31e3077-04e1-4adb-a0cc-44cf84a89eda {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.349928] env[62552]: INFO nova.compute.manager [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] instance snapshotting [ 898.351850] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.362287] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab3f20b-2fc5-4466-b340-9b9f8dff4094 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.395831] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d115f4fb-bdda-4933-93fc-d486440ea4cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.406673] env[62552]: DEBUG oslo_vmware.api [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239631, 'name': PowerOffVM_Task, 'duration_secs': 0.207376} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.406930] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239630, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058762} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.410365] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 898.415498] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Reconfiguring VM instance instance-00000047 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 898.415781] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.416068] env[62552]: DEBUG oslo_vmware.api [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239632, 'name': PowerOnVM_Task, 'duration_secs': 0.479298} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.430762] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14455d00-58ee-4b9f-a6aa-a6f0f5ab5442 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.444082] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b7493b-79b8-4723-bc4e-4f65e19edc9b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.446494] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.446692] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-71356062-acaf-4328-956f-bca060a1bd65 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance '22357d4e-9771-477c-9fc3-fe3d76f6e902' progress to 100 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 898.452314] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d771e1-d5f0-4932-8a8c-0c498b0829e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.456308] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f3b2b7-d345-4271-8d04-12f73ca00cf3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.481562] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 11831580-1b58-476a-91ce-a4e55947fd91/11831580-1b58-476a-91ce-a4e55947fd91.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.493223] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc78bbeb-1e7c-47c8-a397-200210adc3f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.508401] env[62552]: DEBUG nova.compute.provider_tree [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.513019] env[62552]: DEBUG oslo_vmware.api [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 898.513019] env[62552]: value = "task-1239633" [ 898.513019] env[62552]: _type = "Task" [ 898.513019] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.519600] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 898.519600] env[62552]: value = "task-1239634" [ 898.519600] env[62552]: _type = "Task" [ 898.519600] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.523891] env[62552]: DEBUG oslo_vmware.api [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239633, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.532087] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239634, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.849029] env[62552]: DEBUG nova.compute.manager [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 898.852505] env[62552]: DEBUG nova.objects.instance [None req-e7115138-b6f2-47b9-a041-913a90588cc2 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'pci_requests' on Instance uuid e31e3077-04e1-4adb-a0cc-44cf84a89eda {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.878097] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.878097] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.878097] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.878097] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.878097] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.878782] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.879155] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.879466] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.879758] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.880052] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.880341] env[62552]: DEBUG nova.virt.hardware [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.881730] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5312b103-ad1d-4c2c-9ca7-9844360be8e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.890789] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b19cbf7-7781-4a45-948b-30bb1820d03f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.896109] env[62552]: DEBUG nova.network.neutron [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 899.018020] env[62552]: DEBUG nova.scheduler.client.report [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 899.020277] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 899.028065] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2a8402e7-53a8-48d9-ba34-2922e6c6a50d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.043238] env[62552]: DEBUG oslo_vmware.api [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239633, 'name': ReconfigVM_Task, 'duration_secs': 0.382403} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.049579] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Reconfigured VM instance instance-00000047 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 899.050231] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.052022] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239634, 'name': ReconfigVM_Task, 'duration_secs': 0.371269} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.052613] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 899.052613] env[62552]: value = "task-1239635" [ 899.052613] env[62552]: _type = "Task" [ 899.052613] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.054030] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-169a8970-6868-480f-b836-d7bbbd569749 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.055896] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 11831580-1b58-476a-91ce-a4e55947fd91/11831580-1b58-476a-91ce-a4e55947fd91.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.056915] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e547ddc-5387-45f6-ad97-fe4422ea3016 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.071377] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239635, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.073848] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 899.073848] env[62552]: value = "task-1239637" [ 899.073848] env[62552]: _type = "Task" [ 899.073848] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.074107] env[62552]: DEBUG oslo_vmware.api [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 899.074107] env[62552]: value = "task-1239636" [ 899.074107] env[62552]: _type = "Task" [ 899.074107] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.089986] env[62552]: DEBUG oslo_vmware.api [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239636, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.090432] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239637, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.200229] env[62552]: DEBUG nova.network.neutron [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Updating instance_info_cache with network_info: [{"id": "24b537e1-3264-41fe-8165-63c833fc5c62", "address": "fa:16:3e:d8:d9:2c", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24b537e1-32", "ovs_interfaceid": "24b537e1-3264-41fe-8165-63c833fc5c62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.290078] env[62552]: DEBUG nova.network.neutron [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Successfully updated port: a9ece5cd-e814-4895-b895-6ac928e22cfb {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.356911] env[62552]: DEBUG nova.objects.base [None req-e7115138-b6f2-47b9-a041-913a90588cc2 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 899.357152] env[62552]: DEBUG nova.network.neutron [None req-e7115138-b6f2-47b9-a041-913a90588cc2 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 899.381212] env[62552]: DEBUG nova.compute.manager [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Received event network-vif-plugged-24b537e1-3264-41fe-8165-63c833fc5c62 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 899.381459] env[62552]: DEBUG oslo_concurrency.lockutils [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] Acquiring lock "6c8b5367-4fa0-479b-9382-ff261201e3ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.381670] env[62552]: DEBUG oslo_concurrency.lockutils [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] Lock "6c8b5367-4fa0-479b-9382-ff261201e3ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.381843] env[62552]: DEBUG oslo_concurrency.lockutils [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] Lock "6c8b5367-4fa0-479b-9382-ff261201e3ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.382056] env[62552]: DEBUG nova.compute.manager [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] No waiting events found dispatching network-vif-plugged-24b537e1-3264-41fe-8165-63c833fc5c62 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 899.382231] env[62552]: WARNING nova.compute.manager [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Received unexpected event network-vif-plugged-24b537e1-3264-41fe-8165-63c833fc5c62 for instance with vm_state building and task_state spawning. [ 899.382401] env[62552]: DEBUG nova.compute.manager [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Received event network-changed-24b537e1-3264-41fe-8165-63c833fc5c62 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 899.382558] env[62552]: DEBUG nova.compute.manager [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Refreshing instance network info cache due to event network-changed-24b537e1-3264-41fe-8165-63c833fc5c62. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 899.382750] env[62552]: DEBUG oslo_concurrency.lockutils [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] Acquiring lock "refresh_cache-6c8b5367-4fa0-479b-9382-ff261201e3ef" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.468230] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e7115138-b6f2-47b9-a041-913a90588cc2 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.121s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.526245] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.323s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.526763] env[62552]: DEBUG nova.compute.manager [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 899.529489] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.921s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.531533] env[62552]: INFO nova.compute.claims [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.567471] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239635, 'name': CreateSnapshot_Task, 'duration_secs': 0.439135} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.568863] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 899.570284] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab54136-cd40-46f6-a2bd-dac36f5f3e51 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.590840] env[62552]: DEBUG oslo_vmware.api [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239636, 'name': PowerOnVM_Task, 'duration_secs': 0.518632} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.593656] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 899.593932] env[62552]: DEBUG nova.compute.manager [None req-b75ce1ec-e6b0-458f-8be1-0092802be6c2 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 899.594185] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239637, 'name': Rename_Task, 'duration_secs': 0.148645} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.594856] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148d1e31-afd2-4427-898f-1f80cccddc71 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.597190] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.597640] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e508eac-2f34-477c-b376-91eba36cf6c6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.605305] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 899.605305] env[62552]: value = "task-1239638" [ 899.605305] env[62552]: _type = "Task" [ 899.605305] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.611976] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239638, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.703428] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "refresh_cache-6c8b5367-4fa0-479b-9382-ff261201e3ef" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.703834] env[62552]: DEBUG nova.compute.manager [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Instance network_info: |[{"id": "24b537e1-3264-41fe-8165-63c833fc5c62", "address": "fa:16:3e:d8:d9:2c", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24b537e1-32", "ovs_interfaceid": "24b537e1-3264-41fe-8165-63c833fc5c62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 899.704189] env[62552]: DEBUG oslo_concurrency.lockutils [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] Acquired lock "refresh_cache-6c8b5367-4fa0-479b-9382-ff261201e3ef" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.704388] env[62552]: DEBUG nova.network.neutron [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Refreshing network info cache for port 24b537e1-3264-41fe-8165-63c833fc5c62 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 899.705638] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:d9:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '838c9497-35dd-415e-96c7-8dc21b0cd4b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '24b537e1-3264-41fe-8165-63c833fc5c62', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.713278] env[62552]: DEBUG oslo.service.loopingcall [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.713858] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 899.714123] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00d2ff71-c60e-436e-a42c-9ed241826dbd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.736017] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.736017] env[62552]: value = "task-1239639" [ 899.736017] env[62552]: _type = "Task" [ 899.736017] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.744066] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239639, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.793042] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquiring lock "refresh_cache-43d1e732-7e9b-4f9f-b67c-f0ef2be91902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.793673] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquired lock "refresh_cache-43d1e732-7e9b-4f9f-b67c-f0ef2be91902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.793673] env[62552]: DEBUG nova.network.neutron [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.038787] env[62552]: DEBUG nova.compute.utils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.040291] env[62552]: DEBUG nova.compute.manager [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 900.040481] env[62552]: DEBUG nova.network.neutron [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 900.093529] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 900.095307] env[62552]: DEBUG nova.policy [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecfe0d4d977540d99709760db0564cf9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3758acde6e3e4320bf63d4cd1667bfc2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 900.097382] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8569dcb6-7028-45e3-82a9-67d31d45dfa6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.106573] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 900.106573] env[62552]: value = "task-1239640" [ 900.106573] env[62552]: _type = "Task" [ 900.106573] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.122258] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239638, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.125573] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239640, 'name': CloneVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.247254] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239639, 'name': CreateVM_Task, 'duration_secs': 0.327189} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.247474] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.248226] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.248430] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.248782] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.249647] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1f84259-6b19-412f-88ed-d50bd321a588 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.255594] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 900.255594] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f5791f-083e-bd1d-f898-2057a507163f" [ 900.255594] env[62552]: _type = "Task" [ 900.255594] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.264559] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f5791f-083e-bd1d-f898-2057a507163f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.370628] env[62552]: DEBUG nova.network.neutron [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.544660] env[62552]: DEBUG nova.compute.manager [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 900.564742] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "22357d4e-9771-477c-9fc3-fe3d76f6e902" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.565030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.565249] env[62552]: DEBUG nova.compute.manager [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Going to confirm migration 1 {{(pid=62552) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 900.616893] env[62552]: DEBUG oslo_vmware.api [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239638, 'name': PowerOnVM_Task, 'duration_secs': 0.613456} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.619732] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.620047] env[62552]: INFO nova.compute.manager [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Took 9.58 seconds to spawn the instance on the hypervisor. [ 900.620271] env[62552]: DEBUG nova.compute.manager [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 900.621573] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e328ba4-e7ef-4321-874c-f5ced08f3048 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.627198] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239640, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.778095] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f5791f-083e-bd1d-f898-2057a507163f, 'name': SearchDatastore_Task, 'duration_secs': 0.013187} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.778095] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.778095] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.778095] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.778095] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.778095] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.778095] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a82c1da6-95d1-48e2-ac55-437f1918882e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.791732] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.791928] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.797011] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4225f926-7bc3-46e9-a6dc-89c26a8a1000 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.802796] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 900.802796] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5287973b-64d4-45b7-6e33-6907c2da870d" [ 900.802796] env[62552]: _type = "Task" [ 900.802796] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.812683] env[62552]: DEBUG nova.network.neutron [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Updating instance_info_cache with network_info: [{"id": "a9ece5cd-e814-4895-b895-6ac928e22cfb", "address": "fa:16:3e:4b:20:35", "network": {"id": "96474f29-6a91-44c4-8c77-cbfea7fd8785", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-227414207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42f9b9d438854dcd870e6b99cf496307", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9ece5cd-e8", "ovs_interfaceid": "a9ece5cd-e814-4895-b895-6ac928e22cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.820100] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5287973b-64d4-45b7-6e33-6907c2da870d, 'name': SearchDatastore_Task, 'duration_secs': 0.011944} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.820820] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0168b696-8bcc-4514-a0e6-658687b191b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.828276] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 900.828276] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f19652-88d5-ac66-0c33-f6006fb6d6d2" [ 900.828276] env[62552]: _type = "Task" [ 900.828276] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.833958] env[62552]: DEBUG nova.network.neutron [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Updated VIF entry in instance network info cache for port 24b537e1-3264-41fe-8165-63c833fc5c62. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 900.835536] env[62552]: DEBUG nova.network.neutron [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Updating instance_info_cache with network_info: [{"id": "24b537e1-3264-41fe-8165-63c833fc5c62", "address": "fa:16:3e:d8:d9:2c", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24b537e1-32", "ovs_interfaceid": "24b537e1-3264-41fe-8165-63c833fc5c62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.851857] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f19652-88d5-ac66-0c33-f6006fb6d6d2, 'name': SearchDatastore_Task, 'duration_secs': 0.012334} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.852174] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.852478] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 6c8b5367-4fa0-479b-9382-ff261201e3ef/6c8b5367-4fa0-479b-9382-ff261201e3ef.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 900.852736] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a9c3dbe-fc58-4fd7-8133-0d55f0752d34 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.860665] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 900.860665] env[62552]: value = "task-1239641" [ 900.860665] env[62552]: _type = "Task" [ 900.860665] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.874541] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.942448] env[62552]: DEBUG nova.network.neutron [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Successfully created port: 9136fa4c-29b1-49d3-ba65-6530ce9441b8 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.972627] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de93092-4985-4988-bb25-5cfe4e6d317d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.979986] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70825c09-2846-447c-b31b-5610b504d1e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.010055] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d147a20-4f14-4b13-9ead-ba42ae6e8f03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.018069] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b119fee-ce9b-4e02-b700-7d731459b2ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.031328] env[62552]: DEBUG nova.compute.provider_tree [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.120580] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239640, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.142335] env[62552]: INFO nova.compute.manager [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Took 49.97 seconds to build instance. [ 901.164007] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.164658] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.164658] env[62552]: DEBUG nova.network.neutron [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 901.164658] env[62552]: DEBUG nova.objects.instance [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lazy-loading 'info_cache' on Instance uuid 22357d4e-9771-477c-9fc3-fe3d76f6e902 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.316050] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Releasing lock "refresh_cache-43d1e732-7e9b-4f9f-b67c-f0ef2be91902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.316469] env[62552]: DEBUG nova.compute.manager [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Instance network_info: |[{"id": "a9ece5cd-e814-4895-b895-6ac928e22cfb", "address": "fa:16:3e:4b:20:35", "network": {"id": "96474f29-6a91-44c4-8c77-cbfea7fd8785", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-227414207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42f9b9d438854dcd870e6b99cf496307", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9ece5cd-e8", "ovs_interfaceid": "a9ece5cd-e814-4895-b895-6ac928e22cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 901.316987] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:20:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a9ece5cd-e814-4895-b895-6ac928e22cfb', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.326179] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Creating folder: Project (42f9b9d438854dcd870e6b99cf496307). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.326983] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-45e9acd2-91f0-4de5-bf11-c1273bf53616 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.341444] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Created folder: Project (42f9b9d438854dcd870e6b99cf496307) in parent group-v267339. [ 901.341444] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Creating folder: Instances. Parent ref: group-v267460. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.341444] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89e738fb-96b4-40ba-82bd-4565959b8fce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.345660] env[62552]: DEBUG oslo_concurrency.lockutils [req-da9c6528-d540-4756-a40e-c00ca95f3a89 req-ac62b76f-107b-4d30-9cc1-bb40d3ed3951 service nova] Releasing lock "refresh_cache-6c8b5367-4fa0-479b-9382-ff261201e3ef" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.352151] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Created folder: Instances in parent group-v267460. [ 901.352502] env[62552]: DEBUG oslo.service.loopingcall [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.352734] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.353636] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9352b48b-1e1f-4fac-8d02-b81d9f1ecfec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.384856] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239641, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.386667] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.386667] env[62552]: value = "task-1239644" [ 901.386667] env[62552]: _type = "Task" [ 901.386667] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.395461] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239644, 'name': CreateVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.534810] env[62552]: DEBUG nova.scheduler.client.report [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 901.555142] env[62552]: DEBUG nova.compute.manager [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 901.582881] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.583205] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.583427] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.583663] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.583875] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.584089] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.584357] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.584565] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.584891] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.585139] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.585368] env[62552]: DEBUG nova.virt.hardware [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.586267] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9ab583-999d-4a32-9b03-1f50fd5e9fc7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.594613] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf4a05f-aee7-4e86-abc8-21cc2a82dbe5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.619749] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239640, 'name': CloneVM_Task} progress is 95%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.644930] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9a0969fa-260e-475a-9e63-7bf544250ac1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "11831580-1b58-476a-91ce-a4e55947fd91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.477s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.697234] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.697519] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.697848] env[62552]: DEBUG nova.objects.instance [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'flavor' on Instance uuid e31e3077-04e1-4adb-a0cc-44cf84a89eda {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.704266] env[62552]: DEBUG nova.compute.manager [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Received event network-vif-plugged-a9ece5cd-e814-4895-b895-6ac928e22cfb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 901.704266] env[62552]: DEBUG oslo_concurrency.lockutils [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] Acquiring lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.704266] env[62552]: DEBUG oslo_concurrency.lockutils [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] Lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.704266] env[62552]: DEBUG oslo_concurrency.lockutils [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] Lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.704266] env[62552]: DEBUG nova.compute.manager [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] No waiting events found dispatching network-vif-plugged-a9ece5cd-e814-4895-b895-6ac928e22cfb {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 901.704734] env[62552]: WARNING nova.compute.manager [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Received unexpected event network-vif-plugged-a9ece5cd-e814-4895-b895-6ac928e22cfb for instance with vm_state building and task_state spawning. [ 901.704734] env[62552]: DEBUG nova.compute.manager [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Received event network-changed-a9ece5cd-e814-4895-b895-6ac928e22cfb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 901.704734] env[62552]: DEBUG nova.compute.manager [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Refreshing instance network info cache due to event network-changed-a9ece5cd-e814-4895-b895-6ac928e22cfb. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 901.704831] env[62552]: DEBUG oslo_concurrency.lockutils [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] Acquiring lock "refresh_cache-43d1e732-7e9b-4f9f-b67c-f0ef2be91902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.704939] env[62552]: DEBUG oslo_concurrency.lockutils [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] Acquired lock "refresh_cache-43d1e732-7e9b-4f9f-b67c-f0ef2be91902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.705107] env[62552]: DEBUG nova.network.neutron [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Refreshing network info cache for port a9ece5cd-e814-4895-b895-6ac928e22cfb {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.881749] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239641, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.610402} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.882013] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 6c8b5367-4fa0-479b-9382-ff261201e3ef/6c8b5367-4fa0-479b-9382-ff261201e3ef.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 901.882246] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.882492] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c7eb16f9-f5b4-4b62-b38b-041e3b159ee7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.890668] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 901.890668] env[62552]: value = "task-1239645" [ 901.890668] env[62552]: _type = "Task" [ 901.890668] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.897803] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239644, 'name': CreateVM_Task, 'duration_secs': 0.346555} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.898326] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.899044] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.899245] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.899583] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.902532] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b44c3413-55a5-4bbf-b76d-3466f4dbb694 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.904134] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239645, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.906883] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 901.906883] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520ab195-c902-ec3e-82c0-65f6ed2fb392" [ 901.906883] env[62552]: _type = "Task" [ 901.906883] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.914674] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520ab195-c902-ec3e-82c0-65f6ed2fb392, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.930386] env[62552]: DEBUG oslo_concurrency.lockutils [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "5cee4242-49ae-4cb7-a208-e2982f52fbad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.930664] env[62552]: DEBUG oslo_concurrency.lockutils [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "5cee4242-49ae-4cb7-a208-e2982f52fbad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.930940] env[62552]: DEBUG oslo_concurrency.lockutils [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "5cee4242-49ae-4cb7-a208-e2982f52fbad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.931208] env[62552]: DEBUG oslo_concurrency.lockutils [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "5cee4242-49ae-4cb7-a208-e2982f52fbad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.931422] env[62552]: DEBUG oslo_concurrency.lockutils [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "5cee4242-49ae-4cb7-a208-e2982f52fbad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.933680] env[62552]: INFO nova.compute.manager [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Terminating instance [ 902.040992] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.041594] env[62552]: DEBUG nova.compute.manager [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 902.044399] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 30.276s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.044570] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.044719] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 902.045042] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.062s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.046456] env[62552]: INFO nova.compute.claims [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 902.049696] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9266d76-85ea-47b2-8ce6-66088e2d33f3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.057888] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbcddbcb-cbda-45ba-a88c-0ba12dbb315c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.071229] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3baf6afc-6166-4075-8388-b2ea43d8a584 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.077776] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7b4370-e6ba-43df-b1cb-a90567b4739d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.108594] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179650MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 902.108748] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.120832] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239640, 'name': CloneVM_Task, 'duration_secs': 1.747631} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.121416] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Created linked-clone VM from snapshot [ 902.121938] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14efa236-3137-4658-b1f3-b456e5b66c3b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.129152] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Uploading image 9f1bf645-d022-4cd1-9ecf-2fc47981b465 {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 902.149139] env[62552]: DEBUG oslo_vmware.rw_handles [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 902.149139] env[62552]: value = "vm-267459" [ 902.149139] env[62552]: _type = "VirtualMachine" [ 902.149139] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 902.149421] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-91d1e87a-0700-4fd0-b1ec-2d6ff9d32a27 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.156135] env[62552]: DEBUG oslo_vmware.rw_handles [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lease: (returnval){ [ 902.156135] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e57082-8626-aeca-c8f5-1a3417874fc6" [ 902.156135] env[62552]: _type = "HttpNfcLease" [ 902.156135] env[62552]: } obtained for exporting VM: (result){ [ 902.156135] env[62552]: value = "vm-267459" [ 902.156135] env[62552]: _type = "VirtualMachine" [ 902.156135] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 902.156569] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the lease: (returnval){ [ 902.156569] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e57082-8626-aeca-c8f5-1a3417874fc6" [ 902.156569] env[62552]: _type = "HttpNfcLease" [ 902.156569] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 902.162134] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.162134] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e57082-8626-aeca-c8f5-1a3417874fc6" [ 902.162134] env[62552]: _type = "HttpNfcLease" [ 902.162134] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 902.400379] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239645, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063053} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.401208] env[62552]: DEBUG nova.network.neutron [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Updated VIF entry in instance network info cache for port a9ece5cd-e814-4895-b895-6ac928e22cfb. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 902.401605] env[62552]: DEBUG nova.network.neutron [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Updating instance_info_cache with network_info: [{"id": "a9ece5cd-e814-4895-b895-6ac928e22cfb", "address": "fa:16:3e:4b:20:35", "network": {"id": "96474f29-6a91-44c4-8c77-cbfea7fd8785", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-227414207-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42f9b9d438854dcd870e6b99cf496307", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9ece5cd-e8", "ovs_interfaceid": "a9ece5cd-e814-4895-b895-6ac928e22cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.402727] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.403712] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee18949-f810-4bdd-9187-028c559208c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.419767] env[62552]: DEBUG nova.objects.instance [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'pci_requests' on Instance uuid e31e3077-04e1-4adb-a0cc-44cf84a89eda {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.429291] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 6c8b5367-4fa0-479b-9382-ff261201e3ef/6c8b5367-4fa0-479b-9382-ff261201e3ef.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.433079] env[62552]: DEBUG nova.objects.base [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 902.433321] env[62552]: DEBUG nova.network.neutron [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 902.435191] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20d2cad9-a582-491e-b738-ecf36e5d5f2f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.453664] env[62552]: DEBUG nova.compute.manager [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 902.453881] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 902.454894] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520ab195-c902-ec3e-82c0-65f6ed2fb392, 'name': SearchDatastore_Task, 'duration_secs': 0.013545} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.455614] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a049fbc5-cd90-4f5e-a8cd-caa02d9e8ecb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.458632] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.458922] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.459247] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.459422] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.459607] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.460105] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2bac78ee-962f-4a23-8a0b-f415823102f4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.463973] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 902.463973] env[62552]: value = "task-1239647" [ 902.463973] env[62552]: _type = "Task" [ 902.463973] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.469116] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 902.469696] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53609373-c972-4525-90d7-9eebc7f2ed6d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.480912] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239647, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.482109] env[62552]: DEBUG oslo_vmware.api [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 902.482109] env[62552]: value = "task-1239648" [ 902.482109] env[62552]: _type = "Task" [ 902.482109] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.486791] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.486925] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.490702] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b97fe7f0-fe60-45da-a73b-a48612be6531 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.493153] env[62552]: DEBUG oslo_vmware.api [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239648, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.496143] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 902.496143] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a8cb09-13d9-df4a-f3b5-0a8d2b99b24f" [ 902.496143] env[62552]: _type = "Task" [ 902.496143] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.504205] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a8cb09-13d9-df4a-f3b5-0a8d2b99b24f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.555964] env[62552]: DEBUG nova.compute.utils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.555964] env[62552]: DEBUG nova.compute.manager [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 902.555964] env[62552]: DEBUG nova.network.neutron [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 902.583212] env[62552]: DEBUG nova.policy [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6515846d33e44bca942767f9e1d650b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2f57f72de0c4adc9754458f4e7040b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.595508] env[62552]: DEBUG nova.policy [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e321c404f93d4b8986a34ef3b8281e80', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fef66bb035e948079e91a8d9c5786b73', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.664809] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.664809] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e57082-8626-aeca-c8f5-1a3417874fc6" [ 902.664809] env[62552]: _type = "HttpNfcLease" [ 902.664809] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 902.665254] env[62552]: DEBUG oslo_vmware.rw_handles [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 902.665254] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e57082-8626-aeca-c8f5-1a3417874fc6" [ 902.665254] env[62552]: _type = "HttpNfcLease" [ 902.665254] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 902.665954] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b931adff-1914-49f8-aef0-390faf9dabc9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.673199] env[62552]: DEBUG oslo_vmware.rw_handles [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52722003-8448-e6b5-2a48-4c25bdfb3c62/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 902.673507] env[62552]: DEBUG oslo_vmware.rw_handles [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52722003-8448-e6b5-2a48-4c25bdfb3c62/disk-0.vmdk for reading. {{(pid=62552) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 902.733172] env[62552]: DEBUG nova.network.neutron [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance_info_cache with network_info: [{"id": "f749d931-7a86-41ca-a551-4f8d2013603a", "address": "fa:16:3e:00:02:b8", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf749d931-7a", "ovs_interfaceid": "f749d931-7a86-41ca-a551-4f8d2013603a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.769552] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-083f8e54-f82e-475e-a0c9-824ef4e92d77 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.876487] env[62552]: DEBUG nova.network.neutron [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Successfully created port: 3e0578b1-d107-4e54-8f04-1299abff83ef {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.904051] env[62552]: DEBUG oslo_concurrency.lockutils [req-5b0f7d96-caee-4ab0-bd00-fb0779dd4687 req-3a072311-d886-4769-bbda-4a8436b3abc7 service nova] Releasing lock "refresh_cache-43d1e732-7e9b-4f9f-b67c-f0ef2be91902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.976739] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239647, 'name': ReconfigVM_Task, 'duration_secs': 0.278767} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.977173] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 6c8b5367-4fa0-479b-9382-ff261201e3ef/6c8b5367-4fa0-479b-9382-ff261201e3ef.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.978018] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f08a45a-edd9-4cff-8dee-2ef75a41d3dc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.988679] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 902.988679] env[62552]: value = "task-1239649" [ 902.988679] env[62552]: _type = "Task" [ 902.988679] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.998410] env[62552]: DEBUG oslo_vmware.api [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239648, 'name': PowerOffVM_Task, 'duration_secs': 0.516217} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.007868] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.007868] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.012416] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f17efc9-8a6c-473a-b229-73fb11319061 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.013139] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239649, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.019026] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a8cb09-13d9-df4a-f3b5-0a8d2b99b24f, 'name': SearchDatastore_Task, 'duration_secs': 0.013642} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.019847] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-448949ab-c4ec-442d-bee5-70567270707f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.024761] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 903.024761] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bbe088-1308-7a65-3c33-be5c568163b5" [ 903.024761] env[62552]: _type = "Task" [ 903.024761] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.032922] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bbe088-1308-7a65-3c33-be5c568163b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.059339] env[62552]: DEBUG nova.compute.manager [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 903.073857] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.074185] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.074310] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Deleting the datastore file [datastore2] 5cee4242-49ae-4cb7-a208-e2982f52fbad {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.074543] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2dc70680-b7c0-4dac-b887-aa2e1f1b76a9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.082298] env[62552]: DEBUG oslo_vmware.api [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 903.082298] env[62552]: value = "task-1239651" [ 903.082298] env[62552]: _type = "Task" [ 903.082298] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.095032] env[62552]: DEBUG oslo_vmware.api [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.151701] env[62552]: DEBUG nova.network.neutron [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Successfully created port: 035ebcb4-aada-41ce-9361-47d4162965cf {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.238300] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-22357d4e-9771-477c-9fc3-fe3d76f6e902" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.238652] env[62552]: DEBUG nova.objects.instance [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lazy-loading 'migration_context' on Instance uuid 22357d4e-9771-477c-9fc3-fe3d76f6e902 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.295524] env[62552]: DEBUG nova.compute.manager [req-1cb37641-1197-4449-8bb4-26e06ab1d35b req-2800aa14-7505-4b85-b128-b0d90d61a7f8 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Received event network-vif-plugged-9136fa4c-29b1-49d3-ba65-6530ce9441b8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 903.295956] env[62552]: DEBUG oslo_concurrency.lockutils [req-1cb37641-1197-4449-8bb4-26e06ab1d35b req-2800aa14-7505-4b85-b128-b0d90d61a7f8 service nova] Acquiring lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.295956] env[62552]: DEBUG oslo_concurrency.lockutils [req-1cb37641-1197-4449-8bb4-26e06ab1d35b req-2800aa14-7505-4b85-b128-b0d90d61a7f8 service nova] Lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.296412] env[62552]: DEBUG oslo_concurrency.lockutils [req-1cb37641-1197-4449-8bb4-26e06ab1d35b req-2800aa14-7505-4b85-b128-b0d90d61a7f8 service nova] Lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.296997] env[62552]: DEBUG nova.compute.manager [req-1cb37641-1197-4449-8bb4-26e06ab1d35b req-2800aa14-7505-4b85-b128-b0d90d61a7f8 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] No waiting events found dispatching network-vif-plugged-9136fa4c-29b1-49d3-ba65-6530ce9441b8 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 903.297137] env[62552]: WARNING nova.compute.manager [req-1cb37641-1197-4449-8bb4-26e06ab1d35b req-2800aa14-7505-4b85-b128-b0d90d61a7f8 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Received unexpected event network-vif-plugged-9136fa4c-29b1-49d3-ba65-6530ce9441b8 for instance with vm_state building and task_state spawning. [ 903.476360] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f531e5-1090-4be4-997e-2fea6c695df2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.486175] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745a2395-1e98-4d91-a5f7-f7377afee9da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.527649] env[62552]: DEBUG nova.network.neutron [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Successfully updated port: 9136fa4c-29b1-49d3-ba65-6530ce9441b8 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 903.533859] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aacb16a9-549f-4fad-ad7e-26b20bbea34a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.542442] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239649, 'name': Rename_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.547563] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bbe088-1308-7a65-3c33-be5c568163b5, 'name': SearchDatastore_Task, 'duration_secs': 0.00894} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.551694] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.552159] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 43d1e732-7e9b-4f9f-b67c-f0ef2be91902/43d1e732-7e9b-4f9f-b67c-f0ef2be91902.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.552919] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e34768cb-a9ce-4f64-b31c-a7dd4d241b66 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.556157] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79783aec-c590-4cbf-a163-93ae634281b3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.577040] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 903.577040] env[62552]: value = "task-1239652" [ 903.577040] env[62552]: _type = "Task" [ 903.577040] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.577699] env[62552]: DEBUG nova.compute.provider_tree [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.590413] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.594582] env[62552]: DEBUG oslo_vmware.api [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239651, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.309692} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.594963] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.595196] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 903.595414] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 903.595637] env[62552]: INFO nova.compute.manager [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Took 1.14 seconds to destroy the instance on the hypervisor. [ 903.596569] env[62552]: DEBUG oslo.service.loopingcall [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.596569] env[62552]: DEBUG nova.compute.manager [-] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 903.596569] env[62552]: DEBUG nova.network.neutron [-] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 903.744921] env[62552]: DEBUG nova.objects.base [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Object Instance<22357d4e-9771-477c-9fc3-fe3d76f6e902> lazy-loaded attributes: info_cache,migration_context {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 903.745924] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bae9c7-c9e7-46bd-a3dc-de87430258e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.766614] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c4c329d-22e3-4c16-9f28-ab675cc43631 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.772411] env[62552]: DEBUG oslo_vmware.api [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 903.772411] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52482aca-3f74-58cc-c8ff-5962aa72d39e" [ 903.772411] env[62552]: _type = "Task" [ 903.772411] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.780702] env[62552]: DEBUG oslo_vmware.api [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52482aca-3f74-58cc-c8ff-5962aa72d39e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.000485] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239649, 'name': Rename_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.042446] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "refresh_cache-fbbc360d-9cfd-48f1-80b2-26da2c72c002" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.042694] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "refresh_cache-fbbc360d-9cfd-48f1-80b2-26da2c72c002" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.042858] env[62552]: DEBUG nova.network.neutron [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.081478] env[62552]: DEBUG nova.compute.manager [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 904.086270] env[62552]: DEBUG nova.scheduler.client.report [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 904.099719] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239652, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.137287] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.137287] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.139329] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.139913] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.140216] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.140507] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.140843] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.141209] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.141712] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.142054] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.142479] env[62552]: DEBUG nova.virt.hardware [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.143475] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb25bac6-2005-4bf4-8b15-deed476ae45d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.163115] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982cabf7-6f6a-422c-8fae-2f8389ab9fe9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.283696] env[62552]: DEBUG oslo_vmware.api [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52482aca-3f74-58cc-c8ff-5962aa72d39e, 'name': SearchDatastore_Task, 'duration_secs': 0.009191} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.283996] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.421498] env[62552]: DEBUG nova.network.neutron [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Successfully updated port: 3e0578b1-d107-4e54-8f04-1299abff83ef {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.500292] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239649, 'name': Rename_Task, 'duration_secs': 1.145262} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.500684] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.500907] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ca34bed-4bb7-456f-91c0-692b204fc11a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.506618] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 904.506618] env[62552]: value = "task-1239653" [ 904.506618] env[62552]: _type = "Task" [ 904.506618] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.514388] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239653, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.574580] env[62552]: DEBUG nova.compute.manager [req-ef75484e-f86c-49d6-b852-45ac875989d4 req-00962ab5-f314-49b8-bede-b793bf376305 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Received event network-changed-9136fa4c-29b1-49d3-ba65-6530ce9441b8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 904.574929] env[62552]: DEBUG nova.compute.manager [req-ef75484e-f86c-49d6-b852-45ac875989d4 req-00962ab5-f314-49b8-bede-b793bf376305 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Refreshing instance network info cache due to event network-changed-9136fa4c-29b1-49d3-ba65-6530ce9441b8. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 904.575113] env[62552]: DEBUG oslo_concurrency.lockutils [req-ef75484e-f86c-49d6-b852-45ac875989d4 req-00962ab5-f314-49b8-bede-b793bf376305 service nova] Acquiring lock "refresh_cache-fbbc360d-9cfd-48f1-80b2-26da2c72c002" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.577560] env[62552]: DEBUG nova.network.neutron [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 904.594486] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.595119] env[62552]: DEBUG nova.compute.manager [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 904.598079] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239652, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.65566} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.600859] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.484s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.601229] env[62552]: DEBUG nova.objects.instance [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lazy-loading 'resources' on Instance uuid 1cfeedac-f71e-42e4-a04f-8a0462c85907 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.602448] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 43d1e732-7e9b-4f9f-b67c-f0ef2be91902/43d1e732-7e9b-4f9f-b67c-f0ef2be91902.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.602704] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.603310] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd7e7d8d-1c3c-42c8-a073-72cce19c422c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.611721] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 904.611721] env[62552]: value = "task-1239654" [ 904.611721] env[62552]: _type = "Task" [ 904.611721] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.623255] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239654, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.653203] env[62552]: DEBUG nova.network.neutron [-] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.925526] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquiring lock "refresh_cache-0080a6b3-cac0-44f3-a037-3f94f1daa275" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.925914] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquired lock "refresh_cache-0080a6b3-cac0-44f3-a037-3f94f1daa275" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.926229] env[62552]: DEBUG nova.network.neutron [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.954933] env[62552]: DEBUG nova.network.neutron [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Successfully updated port: 035ebcb4-aada-41ce-9361-47d4162965cf {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.972273] env[62552]: DEBUG nova.network.neutron [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Updating instance_info_cache with network_info: [{"id": "9136fa4c-29b1-49d3-ba65-6530ce9441b8", "address": "fa:16:3e:bf:29:cb", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9136fa4c-29", "ovs_interfaceid": "9136fa4c-29b1-49d3-ba65-6530ce9441b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.016820] env[62552]: DEBUG oslo_vmware.api [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239653, 'name': PowerOnVM_Task, 'duration_secs': 0.465272} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.017258] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.017452] env[62552]: INFO nova.compute.manager [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Took 9.41 seconds to spawn the instance on the hypervisor. [ 905.017638] env[62552]: DEBUG nova.compute.manager [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 905.018418] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697fc9ed-871d-4391-861a-cea86b42febf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.101243] env[62552]: DEBUG nova.compute.utils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 905.103235] env[62552]: DEBUG nova.compute.manager [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 905.103465] env[62552]: DEBUG nova.network.neutron [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 905.124045] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239654, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069974} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.125931] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.127186] env[62552]: DEBUG nova.compute.manager [req-a42ffa45-4a62-4309-86be-2c21ad05df1c req-1a665fd1-7e44-4705-86d6-8f342a596dfb service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Received event network-vif-plugged-3e0578b1-d107-4e54-8f04-1299abff83ef {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 905.127586] env[62552]: DEBUG oslo_concurrency.lockutils [req-a42ffa45-4a62-4309-86be-2c21ad05df1c req-1a665fd1-7e44-4705-86d6-8f342a596dfb service nova] Acquiring lock "0080a6b3-cac0-44f3-a037-3f94f1daa275-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.127640] env[62552]: DEBUG oslo_concurrency.lockutils [req-a42ffa45-4a62-4309-86be-2c21ad05df1c req-1a665fd1-7e44-4705-86d6-8f342a596dfb service nova] Lock "0080a6b3-cac0-44f3-a037-3f94f1daa275-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.127768] env[62552]: DEBUG oslo_concurrency.lockutils [req-a42ffa45-4a62-4309-86be-2c21ad05df1c req-1a665fd1-7e44-4705-86d6-8f342a596dfb service nova] Lock "0080a6b3-cac0-44f3-a037-3f94f1daa275-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.127938] env[62552]: DEBUG nova.compute.manager [req-a42ffa45-4a62-4309-86be-2c21ad05df1c req-1a665fd1-7e44-4705-86d6-8f342a596dfb service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] No waiting events found dispatching network-vif-plugged-3e0578b1-d107-4e54-8f04-1299abff83ef {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 905.128121] env[62552]: WARNING nova.compute.manager [req-a42ffa45-4a62-4309-86be-2c21ad05df1c req-1a665fd1-7e44-4705-86d6-8f342a596dfb service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Received unexpected event network-vif-plugged-3e0578b1-d107-4e54-8f04-1299abff83ef for instance with vm_state building and task_state spawning. [ 905.128949] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4368a4b-2186-4c06-9f93-55483a223565 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.160864] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 43d1e732-7e9b-4f9f-b67c-f0ef2be91902/43d1e732-7e9b-4f9f-b67c-f0ef2be91902.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.164808] env[62552]: DEBUG nova.policy [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb432303c2394f8795672ceaf02f28aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a0bc011d6794602b2bbe1fc01e4c8b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 905.166393] env[62552]: INFO nova.compute.manager [-] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Took 1.57 seconds to deallocate network for instance. [ 905.166931] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00d4703c-d173-423f-b04f-12b0b5dc68b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.192162] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 905.192162] env[62552]: value = "task-1239655" [ 905.192162] env[62552]: _type = "Task" [ 905.192162] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.204362] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239655, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.460867] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.461191] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.461457] env[62552]: DEBUG nova.network.neutron [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.477925] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "refresh_cache-fbbc360d-9cfd-48f1-80b2-26da2c72c002" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.478285] env[62552]: DEBUG nova.compute.manager [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Instance network_info: |[{"id": "9136fa4c-29b1-49d3-ba65-6530ce9441b8", "address": "fa:16:3e:bf:29:cb", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9136fa4c-29", "ovs_interfaceid": "9136fa4c-29b1-49d3-ba65-6530ce9441b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 905.479346] env[62552]: DEBUG oslo_concurrency.lockutils [req-ef75484e-f86c-49d6-b852-45ac875989d4 req-00962ab5-f314-49b8-bede-b793bf376305 service nova] Acquired lock "refresh_cache-fbbc360d-9cfd-48f1-80b2-26da2c72c002" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.482237] env[62552]: DEBUG nova.network.neutron [req-ef75484e-f86c-49d6-b852-45ac875989d4 req-00962ab5-f314-49b8-bede-b793bf376305 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Refreshing network info cache for port 9136fa4c-29b1-49d3-ba65-6530ce9441b8 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 905.482237] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:29:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9136fa4c-29b1-49d3-ba65-6530ce9441b8', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.491422] env[62552]: DEBUG oslo.service.loopingcall [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.492158] env[62552]: DEBUG nova.network.neutron [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 905.495027] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.497283] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e04ac132-6014-4aee-8237-5a161c59dbee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.521751] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.521751] env[62552]: value = "task-1239656" [ 905.521751] env[62552]: _type = "Task" [ 905.521751] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.535152] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239656, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.539028] env[62552]: DEBUG nova.network.neutron [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Successfully created port: 5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.544214] env[62552]: INFO nova.compute.manager [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Took 52.01 seconds to build instance. [ 905.567511] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1664f60b-5faa-41a3-bdc2-d6216df6ab3b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.579220] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cf3c92-55e7-4feb-bbba-76562ea71f65 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.611827] env[62552]: DEBUG nova.compute.manager [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 905.618108] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f23f2e4-5025-43b5-9dcb-5385a3a8442f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.628725] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409952b0-fa43-4742-a12b-191c7c5460e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.643015] env[62552]: DEBUG nova.compute.provider_tree [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.688630] env[62552]: DEBUG oslo_concurrency.lockutils [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.704712] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239655, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.715735] env[62552]: DEBUG nova.network.neutron [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Updating instance_info_cache with network_info: [{"id": "3e0578b1-d107-4e54-8f04-1299abff83ef", "address": "fa:16:3e:41:0e:e0", "network": {"id": "ded16e05-d290-4c6c-a47c-e1b2e89c7a57", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1330766044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fef66bb035e948079e91a8d9c5786b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e0578b1-d1", "ovs_interfaceid": "3e0578b1-d107-4e54-8f04-1299abff83ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.017850] env[62552]: WARNING nova.network.neutron [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] 06f1ea1a-5103-419f-a356-3b65b9a2685e already exists in list: networks containing: ['06f1ea1a-5103-419f-a356-3b65b9a2685e']. ignoring it [ 906.032063] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239656, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.049841] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d2ef4805-166a-4dfc-8e37-3268f420d5f4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "6c8b5367-4fa0-479b-9382-ff261201e3ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.614s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.148889] env[62552]: DEBUG nova.scheduler.client.report [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 906.205810] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239655, 'name': ReconfigVM_Task, 'duration_secs': 0.585684} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.206146] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 43d1e732-7e9b-4f9f-b67c-f0ef2be91902/43d1e732-7e9b-4f9f-b67c-f0ef2be91902.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.206840] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e9146db-8be1-404c-810c-e78be3961ca2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.213208] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 906.213208] env[62552]: value = "task-1239657" [ 906.213208] env[62552]: _type = "Task" [ 906.213208] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.218556] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Releasing lock "refresh_cache-0080a6b3-cac0-44f3-a037-3f94f1daa275" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.218834] env[62552]: DEBUG nova.compute.manager [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Instance network_info: |[{"id": "3e0578b1-d107-4e54-8f04-1299abff83ef", "address": "fa:16:3e:41:0e:e0", "network": {"id": "ded16e05-d290-4c6c-a47c-e1b2e89c7a57", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1330766044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fef66bb035e948079e91a8d9c5786b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e0578b1-d1", "ovs_interfaceid": "3e0578b1-d107-4e54-8f04-1299abff83ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 906.219840] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:0e:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4a8fd90-153b-494f-b76a-299eb05c03f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e0578b1-d107-4e54-8f04-1299abff83ef', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.227923] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Creating folder: Project (fef66bb035e948079e91a8d9c5786b73). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 906.232042] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7d74cafc-7771-4243-8889-081c15d8595c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.233626] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239657, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.241782] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Created folder: Project (fef66bb035e948079e91a8d9c5786b73) in parent group-v267339. [ 906.242434] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Creating folder: Instances. Parent ref: group-v267464. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 906.242434] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7e724666-d5d9-47d0-a097-a355a0ce7c57 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.251025] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Created folder: Instances in parent group-v267464. [ 906.251330] env[62552]: DEBUG oslo.service.loopingcall [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.251577] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 906.251792] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c54e7ca6-b8e5-47f2-9932-dd8787c8ee63 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.270828] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.270828] env[62552]: value = "task-1239660" [ 906.270828] env[62552]: _type = "Task" [ 906.270828] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.278640] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239660, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.534440] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239656, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.589983] env[62552]: DEBUG nova.network.neutron [req-ef75484e-f86c-49d6-b852-45ac875989d4 req-00962ab5-f314-49b8-bede-b793bf376305 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Updated VIF entry in instance network info cache for port 9136fa4c-29b1-49d3-ba65-6530ce9441b8. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.589983] env[62552]: DEBUG nova.network.neutron [req-ef75484e-f86c-49d6-b852-45ac875989d4 req-00962ab5-f314-49b8-bede-b793bf376305 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Updating instance_info_cache with network_info: [{"id": "9136fa4c-29b1-49d3-ba65-6530ce9441b8", "address": "fa:16:3e:bf:29:cb", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9136fa4c-29", "ovs_interfaceid": "9136fa4c-29b1-49d3-ba65-6530ce9441b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.627527] env[62552]: DEBUG nova.compute.manager [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 906.653933] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.654309] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.654535] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.655444] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.655639] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.655819] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.656114] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.656346] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.656633] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.656849] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.657101] env[62552]: DEBUG nova.virt.hardware [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.658174] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.057s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.661038] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e5b6e9-c466-4e76-884f-07bfc8eb3cc0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.664858] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.078s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.665237] env[62552]: DEBUG nova.objects.instance [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'pci_requests' on Instance uuid cdee22a9-4327-47af-ab14-8403a06b2802 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.674079] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2471ecb2-057d-42d3-b692-dbb859d9c50e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.690728] env[62552]: INFO nova.scheduler.client.report [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Deleted allocations for instance 1cfeedac-f71e-42e4-a04f-8a0462c85907 [ 906.718697] env[62552]: DEBUG nova.network.neutron [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [{"id": "0679d99d-6c79-408f-82a1-9f80978a1c48", "address": "fa:16:3e:c1:12:aa", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0679d99d-6c", "ovs_interfaceid": "0679d99d-6c79-408f-82a1-9f80978a1c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "035ebcb4-aada-41ce-9361-47d4162965cf", "address": "fa:16:3e:5c:87:4a", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap035ebcb4-aa", "ovs_interfaceid": "035ebcb4-aada-41ce-9361-47d4162965cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.726314] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239657, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.781576] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239660, 'name': CreateVM_Task, 'duration_secs': 0.392478} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.781755] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.782659] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.782918] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.783303] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 906.783666] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f8a4c51-691a-426e-beab-f6955fdd8039 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.790277] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 906.790277] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5284a0f4-0d8f-971f-e8ef-fa836e224115" [ 906.790277] env[62552]: _type = "Task" [ 906.790277] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.804455] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5284a0f4-0d8f-971f-e8ef-fa836e224115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.033599] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239656, 'name': CreateVM_Task, 'duration_secs': 1.487634} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.033935] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 907.034773] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.085791] env[62552]: DEBUG nova.network.neutron [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Successfully updated port: 5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 907.091759] env[62552]: DEBUG oslo_concurrency.lockutils [req-ef75484e-f86c-49d6-b852-45ac875989d4 req-00962ab5-f314-49b8-bede-b793bf376305 service nova] Releasing lock "refresh_cache-fbbc360d-9cfd-48f1-80b2-26da2c72c002" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.175518] env[62552]: DEBUG nova.objects.instance [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'numa_topology' on Instance uuid cdee22a9-4327-47af-ab14-8403a06b2802 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.189702] env[62552]: DEBUG nova.compute.manager [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Received event network-vif-deleted-cf27f46e-a6d0-4547-9eea-2a88fb1f21dc {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 907.189929] env[62552]: DEBUG nova.compute.manager [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-vif-plugged-035ebcb4-aada-41ce-9361-47d4162965cf {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 907.190140] env[62552]: DEBUG oslo_concurrency.lockutils [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.190387] env[62552]: DEBUG oslo_concurrency.lockutils [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.190586] env[62552]: DEBUG oslo_concurrency.lockutils [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.190780] env[62552]: DEBUG nova.compute.manager [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] No waiting events found dispatching network-vif-plugged-035ebcb4-aada-41ce-9361-47d4162965cf {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 907.190956] env[62552]: WARNING nova.compute.manager [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received unexpected event network-vif-plugged-035ebcb4-aada-41ce-9361-47d4162965cf for instance with vm_state active and task_state None. [ 907.191279] env[62552]: DEBUG nova.compute.manager [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-changed-035ebcb4-aada-41ce-9361-47d4162965cf {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 907.191454] env[62552]: DEBUG nova.compute.manager [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Refreshing instance network info cache due to event network-changed-035ebcb4-aada-41ce-9361-47d4162965cf. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 907.191651] env[62552]: DEBUG oslo_concurrency.lockutils [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] Acquiring lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.201199] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7810ba15-63ab-4cc9-a4b4-9a74ff573791 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050 tempest-FloatingIPsAssociationNegativeTestJSON-2026447050-project-member] Lock "1cfeedac-f71e-42e4-a04f-8a0462c85907" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.108s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.221480] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.222194] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.222429] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.225951] env[62552]: DEBUG oslo_concurrency.lockutils [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] Acquired lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.226158] env[62552]: DEBUG nova.network.neutron [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Refreshing network info cache for port 035ebcb4-aada-41ce-9361-47d4162965cf {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.228012] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56654391-ae47-4700-91d7-1f34280f2452 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.231252] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239657, 'name': Rename_Task, 'duration_secs': 0.87938} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.232093] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.232715] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de061617-b681-47d3-9674-d8ce21718c99 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.247035] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.247304] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.247469] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.247654] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.247819] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.248049] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.248273] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.248488] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.248600] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.248785] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.248973] env[62552]: DEBUG nova.virt.hardware [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.255388] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Reconfiguring VM to attach interface {{(pid=62552) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 907.257244] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6b58b07-962b-4a2a-83cc-8cffad35ba15 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.270589] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 907.270589] env[62552]: value = "task-1239661" [ 907.270589] env[62552]: _type = "Task" [ 907.270589] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.276511] env[62552]: DEBUG oslo_vmware.api [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 907.276511] env[62552]: value = "task-1239662" [ 907.276511] env[62552]: _type = "Task" [ 907.276511] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.283239] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239661, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.288256] env[62552]: DEBUG oslo_vmware.api [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239662, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.300624] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5284a0f4-0d8f-971f-e8ef-fa836e224115, 'name': SearchDatastore_Task, 'duration_secs': 0.029692} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.300960] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.301310] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.301564] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.301710] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.301897] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.302258] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.302575] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 907.302822] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b94d73b-5f68-4964-80b6-4e32e08eb731 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.304905] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a600ee1-6ca9-435f-9ed0-ab79676374b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.309760] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 907.309760] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52029537-ca8c-9a23-8202-e7d5b76cd8f7" [ 907.309760] env[62552]: _type = "Task" [ 907.309760] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.315343] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.315343] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.319521] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28d5d3ae-6380-4e20-940f-4d56770a1a1d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.322043] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52029537-ca8c-9a23-8202-e7d5b76cd8f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.325455] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 907.325455] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ecf965-8b30-55eb-b06b-9f7bea03e3f0" [ 907.325455] env[62552]: _type = "Task" [ 907.325455] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.337027] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ecf965-8b30-55eb-b06b-9f7bea03e3f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.589475] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.589475] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.589757] env[62552]: DEBUG nova.network.neutron [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.678591] env[62552]: INFO nova.compute.claims [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.787668] env[62552]: DEBUG nova.compute.manager [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Received event network-changed-3e0578b1-d107-4e54-8f04-1299abff83ef {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 907.788165] env[62552]: DEBUG nova.compute.manager [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Refreshing instance network info cache due to event network-changed-3e0578b1-d107-4e54-8f04-1299abff83ef. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 907.788584] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] Acquiring lock "refresh_cache-0080a6b3-cac0-44f3-a037-3f94f1daa275" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.788903] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] Acquired lock "refresh_cache-0080a6b3-cac0-44f3-a037-3f94f1daa275" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.791019] env[62552]: DEBUG nova.network.neutron [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Refreshing network info cache for port 3e0578b1-d107-4e54-8f04-1299abff83ef {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.796330] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239661, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.801679] env[62552]: DEBUG oslo_vmware.api [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239662, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.821361] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52029537-ca8c-9a23-8202-e7d5b76cd8f7, 'name': SearchDatastore_Task, 'duration_secs': 0.01649} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.822239] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.822625] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.823073] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.836324] env[62552]: INFO nova.compute.manager [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Rescuing [ 907.839021] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "refresh_cache-6c8b5367-4fa0-479b-9382-ff261201e3ef" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.839021] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "refresh_cache-6c8b5367-4fa0-479b-9382-ff261201e3ef" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.839021] env[62552]: DEBUG nova.network.neutron [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.839021] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ecf965-8b30-55eb-b06b-9f7bea03e3f0, 'name': SearchDatastore_Task, 'duration_secs': 0.025687} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.840898] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7961e3f5-ba5e-478f-9d78-ab9795a71a95 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.847930] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 907.847930] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52df0670-afbe-9ee6-45dd-9f6f9694156c" [ 907.847930] env[62552]: _type = "Task" [ 907.847930] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.861914] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52df0670-afbe-9ee6-45dd-9f6f9694156c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.122475] env[62552]: DEBUG nova.network.neutron [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updated VIF entry in instance network info cache for port 035ebcb4-aada-41ce-9361-47d4162965cf. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.123479] env[62552]: DEBUG nova.network.neutron [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [{"id": "0679d99d-6c79-408f-82a1-9f80978a1c48", "address": "fa:16:3e:c1:12:aa", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0679d99d-6c", "ovs_interfaceid": "0679d99d-6c79-408f-82a1-9f80978a1c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "035ebcb4-aada-41ce-9361-47d4162965cf", "address": "fa:16:3e:5c:87:4a", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap035ebcb4-aa", "ovs_interfaceid": "035ebcb4-aada-41ce-9361-47d4162965cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.132768] env[62552]: DEBUG nova.network.neutron [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 908.289352] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239661, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.293440] env[62552]: DEBUG oslo_vmware.api [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239662, 'name': ReconfigVM_Task, 'duration_secs': 0.772461} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.294031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.294474] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Reconfigured VM to attach interface {{(pid=62552) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 908.345730] env[62552]: DEBUG nova.network.neutron [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updating instance_info_cache with network_info: [{"id": "5e69a23a-a63b-4d24-9729-859b537f945b", "address": "fa:16:3e:0b:c1:98", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e69a23a-a6", "ovs_interfaceid": "5e69a23a-a63b-4d24-9729-859b537f945b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.361930] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52df0670-afbe-9ee6-45dd-9f6f9694156c, 'name': SearchDatastore_Task, 'duration_secs': 0.016408} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.362103] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.362413] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 0080a6b3-cac0-44f3-a037-3f94f1daa275/0080a6b3-cac0-44f3-a037-3f94f1daa275.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.363375] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.363640] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.363891] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aafc0d1f-edc1-4b6f-8fb8-e0142d1f1b7e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.367839] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b0312b6-f766-4d25-809c-81f20d7ad59a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.375760] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 908.375760] env[62552]: value = "task-1239663" [ 908.375760] env[62552]: _type = "Task" [ 908.375760] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.380063] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.380532] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 908.384076] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce54acc0-a7a6-4a39-881c-ce91e82d6a28 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.390914] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239663, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.394703] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 908.394703] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bc2295-d1ab-29c3-8aaf-f889c3245c29" [ 908.394703] env[62552]: _type = "Task" [ 908.394703] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.403020] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bc2295-d1ab-29c3-8aaf-f889c3245c29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.626495] env[62552]: DEBUG oslo_concurrency.lockutils [req-832ccc26-cf41-43b8-aeb2-77d620bb837b req-191ba5a6-bf8f-4aa6-8a04-f03ce97aaaa8 service nova] Releasing lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.640010] env[62552]: DEBUG nova.network.neutron [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Updating instance_info_cache with network_info: [{"id": "24b537e1-3264-41fe-8165-63c833fc5c62", "address": "fa:16:3e:d8:d9:2c", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24b537e1-32", "ovs_interfaceid": "24b537e1-3264-41fe-8165-63c833fc5c62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.671622] env[62552]: DEBUG nova.network.neutron [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Updated VIF entry in instance network info cache for port 3e0578b1-d107-4e54-8f04-1299abff83ef. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.671673] env[62552]: DEBUG nova.network.neutron [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Updating instance_info_cache with network_info: [{"id": "3e0578b1-d107-4e54-8f04-1299abff83ef", "address": "fa:16:3e:41:0e:e0", "network": {"id": "ded16e05-d290-4c6c-a47c-e1b2e89c7a57", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1330766044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fef66bb035e948079e91a8d9c5786b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e0578b1-d1", "ovs_interfaceid": "3e0578b1-d107-4e54-8f04-1299abff83ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.790251] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239661, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.804150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7990f346-46a9-419e-968e-65879db3afca tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.106s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.851606] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.851606] env[62552]: DEBUG nova.compute.manager [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Instance network_info: |[{"id": "5e69a23a-a63b-4d24-9729-859b537f945b", "address": "fa:16:3e:0b:c1:98", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e69a23a-a6", "ovs_interfaceid": "5e69a23a-a63b-4d24-9729-859b537f945b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 908.852527] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:c1:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e69a23a-a63b-4d24-9729-859b537f945b', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.861954] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Creating folder: Project (1a0bc011d6794602b2bbe1fc01e4c8b0). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 908.862169] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-784c99ef-1cce-48f4-8190-c50a5e21a62a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.881261] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Created folder: Project (1a0bc011d6794602b2bbe1fc01e4c8b0) in parent group-v267339. [ 908.881466] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Creating folder: Instances. Parent ref: group-v267467. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 908.885711] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a38707f-2211-4236-a62c-414e84c5d888 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.894457] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239663, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.907083] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Created folder: Instances in parent group-v267467. [ 908.907457] env[62552]: DEBUG oslo.service.loopingcall [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.908360] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.908637] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-249713cc-3a5e-4c44-ae60-64e4c9f3fb02 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.935155] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bc2295-d1ab-29c3-8aaf-f889c3245c29, 'name': SearchDatastore_Task, 'duration_secs': 0.0121} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.936639] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ba44339-6e1c-4fd0-8854-0b392f3e5d9a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.942368] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.942368] env[62552]: value = "task-1239666" [ 908.942368] env[62552]: _type = "Task" [ 908.942368] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.947138] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 908.947138] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524742e8-e057-aa35-cef9-114ec282aae6" [ 908.947138] env[62552]: _type = "Task" [ 908.947138] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.956953] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239666, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.963260] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524742e8-e057-aa35-cef9-114ec282aae6, 'name': SearchDatastore_Task, 'duration_secs': 0.012738} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.963609] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.964091] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] fbbc360d-9cfd-48f1-80b2-26da2c72c002/fbbc360d-9cfd-48f1-80b2-26da2c72c002.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.964561] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8c5985b-5998-4b89-981d-8d95096072ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.973819] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 908.973819] env[62552]: value = "task-1239667" [ 908.973819] env[62552]: _type = "Task" [ 908.973819] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.986537] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239667, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.133408] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065a5d6a-d738-4b9a-8b64-9d48e21102b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.141357] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01efb307-d160-42e0-8c84-8caa9442adcd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.145256] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "refresh_cache-6c8b5367-4fa0-479b-9382-ff261201e3ef" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.177498] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] Releasing lock "refresh_cache-0080a6b3-cac0-44f3-a037-3f94f1daa275" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.177827] env[62552]: DEBUG nova.compute.manager [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Received event network-vif-plugged-5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 909.178047] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] Acquiring lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.178268] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.178437] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.178609] env[62552]: DEBUG nova.compute.manager [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] No waiting events found dispatching network-vif-plugged-5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 909.178791] env[62552]: WARNING nova.compute.manager [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Received unexpected event network-vif-plugged-5e69a23a-a63b-4d24-9729-859b537f945b for instance with vm_state building and task_state spawning. [ 909.178978] env[62552]: DEBUG nova.compute.manager [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Received event network-changed-5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 909.179186] env[62552]: DEBUG nova.compute.manager [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Refreshing instance network info cache due to event network-changed-5e69a23a-a63b-4d24-9729-859b537f945b. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 909.179384] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] Acquiring lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.179530] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] Acquired lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.179692] env[62552]: DEBUG nova.network.neutron [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Refreshing network info cache for port 5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 909.183538] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4eed138-1b33-447d-a8c8-1c3a2f0b6292 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.194969] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60fa3c64-082f-4ecb-8d07-2f87122d8489 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.211016] env[62552]: DEBUG nova.compute.provider_tree [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.286957] env[62552]: DEBUG oslo_vmware.api [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239661, 'name': PowerOnVM_Task, 'duration_secs': 1.769499} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.287356] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.287584] env[62552]: INFO nova.compute.manager [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Took 10.44 seconds to spawn the instance on the hypervisor. [ 909.287790] env[62552]: DEBUG nova.compute.manager [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 909.288616] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871be3ca-77cd-4ccd-8add-4a823935bb01 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.385637] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239663, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.451984] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239666, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.485698] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239667, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.714363] env[62552]: DEBUG nova.scheduler.client.report [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 909.809963] env[62552]: INFO nova.compute.manager [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Took 50.45 seconds to build instance. [ 909.887983] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239663, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.904841] env[62552]: DEBUG nova.network.neutron [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updated VIF entry in instance network info cache for port 5e69a23a-a63b-4d24-9729-859b537f945b. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 909.905253] env[62552]: DEBUG nova.network.neutron [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updating instance_info_cache with network_info: [{"id": "5e69a23a-a63b-4d24-9729-859b537f945b", "address": "fa:16:3e:0b:c1:98", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e69a23a-a6", "ovs_interfaceid": "5e69a23a-a63b-4d24-9729-859b537f945b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.958231] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239666, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.987107] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239667, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.208733] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.208733] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-697a209b-b418-432b-beee-16c5a8a0d5e4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.214457] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 910.214457] env[62552]: value = "task-1239668" [ 910.214457] env[62552]: _type = "Task" [ 910.214457] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.221563] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.557s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.228630] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 30.348s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.232033] env[62552]: DEBUG nova.objects.instance [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62552) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 910.232562] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.301649] env[62552]: INFO nova.network.neutron [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating port 929d55b0-5a35-4d2d-a172-15d79a215977 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 910.311664] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7b0eef68-f4f6-448c-a8d8-dfff4d4d1501 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.154s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.395884] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239663, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.549621} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.395884] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 0080a6b3-cac0-44f3-a037-3f94f1daa275/0080a6b3-cac0-44f3-a037-3f94f1daa275.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 910.395884] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 910.395884] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc305ba0-2a19-4caf-9ab4-6742d5bf1802 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.402688] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 910.402688] env[62552]: value = "task-1239669" [ 910.402688] env[62552]: _type = "Task" [ 910.402688] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.411935] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9489fe5-afb9-424d-8ca9-7014adfaf019 req-1f3c44c0-d882-4ebd-bc18-4cab0e32d7a8 service nova] Releasing lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.412813] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239669, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.453498] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239666, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.495309] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239667, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.730948] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239668, 'name': PowerOffVM_Task, 'duration_secs': 0.317895} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.731642] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.732872] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9a537a-69fa-40d4-a8f7-9d23bfc1d174 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.756501] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aac20d1-3466-497a-b816-e76ac983ff2b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.791347] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.793625] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01d0db61-7069-4996-8e3a-3bafa560fd9f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.805922] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 910.805922] env[62552]: value = "task-1239670" [ 910.805922] env[62552]: _type = "Task" [ 910.805922] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.816013] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 910.816013] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.816210] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.816381] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.816574] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.816877] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-501d3b38-96de-4ced-80ba-60cc3eb60b19 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.825709] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.825903] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 910.826679] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3d7eda3-b150-4081-9de3-b9e2a72ef47e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.832600] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 910.832600] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524de6e7-229d-09c5-0b9a-2d7469ba857b" [ 910.832600] env[62552]: _type = "Task" [ 910.832600] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.843329] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524de6e7-229d-09c5-0b9a-2d7469ba857b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.911617] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239669, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09718} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.911901] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 910.912731] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01205da-af18-40c9-af20-d1efb3bfddf5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.935987] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 0080a6b3-cac0-44f3-a037-3f94f1daa275/0080a6b3-cac0-44f3-a037-3f94f1daa275.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.936361] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3501407-3fb8-405f-8181-e77a41ea2c55 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.960437] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239666, 'name': CreateVM_Task, 'duration_secs': 1.724002} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.961678] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.962048] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 910.962048] env[62552]: value = "task-1239671" [ 910.962048] env[62552]: _type = "Task" [ 910.962048] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.962752] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.962966] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.963328] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.963689] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6abc0f3e-bcc3-461e-b23f-98dbec44ba59 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.971871] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 910.971871] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5283a306-34d0-e98b-1c3c-8e4af2949434" [ 910.971871] env[62552]: _type = "Task" [ 910.971871] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.975178] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239671, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.987638] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5283a306-34d0-e98b-1c3c-8e4af2949434, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.995538] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239667, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.565085} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.996086] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] fbbc360d-9cfd-48f1-80b2-26da2c72c002/fbbc360d-9cfd-48f1-80b2-26da2c72c002.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 910.996322] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 910.996586] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a50949c8-c9e1-461a-a70b-993f23480477 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.005894] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 911.005894] env[62552]: value = "task-1239672" [ 911.005894] env[62552]: _type = "Task" [ 911.005894] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.014442] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239672, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.219787] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-9726e936-f144-4cce-9964-78bcae1c40ae" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.219787] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-9726e936-f144-4cce-9964-78bcae1c40ae" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.219787] env[62552]: DEBUG nova.objects.instance [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'flavor' on Instance uuid e31e3077-04e1-4adb-a0cc-44cf84a89eda {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.241477] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0d44db7-73cd-451d-a9ec-8d050f08261d tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.242888] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.959s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.244623] env[62552]: INFO nova.compute.claims [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.346888] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524de6e7-229d-09c5-0b9a-2d7469ba857b, 'name': SearchDatastore_Task, 'duration_secs': 0.010523} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.347733] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1267717b-078f-4ea9-b503-ba6b73880b43 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.356021] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 911.356021] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5248e969-806d-7c91-0ba9-f478760db38e" [ 911.356021] env[62552]: _type = "Task" [ 911.356021] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.363313] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5248e969-806d-7c91-0ba9-f478760db38e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.474345] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239671, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.485296] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5283a306-34d0-e98b-1c3c-8e4af2949434, 'name': SearchDatastore_Task, 'duration_secs': 0.014772} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.485681] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.485986] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.486299] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.486842] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.486842] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.487010] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e091f9a-67c5-46ab-8e5e-b4a99ae6c0ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.496523] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquiring lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.496833] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.497077] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquiring lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.497289] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.497590] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.499414] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.499887] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.500983] env[62552]: INFO nova.compute.manager [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Terminating instance [ 911.502571] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27c29c54-6a30-4467-b743-70762b0868d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.518263] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 911.518263] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d0ab50-e38d-1f41-4c80-5039a6a48c2f" [ 911.518263] env[62552]: _type = "Task" [ 911.518263] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.525944] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239672, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078912} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.526766] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.528009] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9bd05e2-01f4-4372-9bf3-e7c5b723f26d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.534307] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d0ab50-e38d-1f41-4c80-5039a6a48c2f, 'name': SearchDatastore_Task, 'duration_secs': 0.012606} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.543790] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0523a49d-802f-4361-886e-52e895a45585 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.561656] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] fbbc360d-9cfd-48f1-80b2-26da2c72c002/fbbc360d-9cfd-48f1-80b2-26da2c72c002.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.562559] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80a4cd35-24c6-4c9b-8f0a-47c80b291e93 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.580512] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 911.580512] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521170c9-8185-c172-28c8-d6c001ddf3c1" [ 911.580512] env[62552]: _type = "Task" [ 911.580512] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.587821] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 911.587821] env[62552]: value = "task-1239673" [ 911.587821] env[62552]: _type = "Task" [ 911.587821] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.596431] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521170c9-8185-c172-28c8-d6c001ddf3c1, 'name': SearchDatastore_Task, 'duration_secs': 0.014106} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.596431] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.596431] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6a5b5f4e-0c84-447d-a3da-8258512abb7c/6a5b5f4e-0c84-447d-a3da-8258512abb7c.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 911.596431] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d491ca5-7477-4ced-96cf-aa9ef7457ca4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.603270] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239673, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.612503] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 911.612503] env[62552]: value = "task-1239674" [ 911.612503] env[62552]: _type = "Task" [ 911.612503] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.622524] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239674, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.867149] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5248e969-806d-7c91-0ba9-f478760db38e, 'name': SearchDatastore_Task, 'duration_secs': 0.012565} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.868083] env[62552]: DEBUG oslo_concurrency.lockutils [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.868083] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 6c8b5367-4fa0-479b-9382-ff261201e3ef/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. {{(pid=62552) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 911.868470] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b91f6bd2-e59a-481f-a14d-a0fdd60fbc71 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.876212] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 911.876212] env[62552]: value = "task-1239675" [ 911.876212] env[62552]: _type = "Task" [ 911.876212] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.888470] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239675, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.895738] env[62552]: DEBUG nova.objects.instance [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'pci_requests' on Instance uuid e31e3077-04e1-4adb-a0cc-44cf84a89eda {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.977233] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239671, 'name': ReconfigVM_Task, 'duration_secs': 0.852406} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.981169] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 0080a6b3-cac0-44f3-a037-3f94f1daa275/0080a6b3-cac0-44f3-a037-3f94f1daa275.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.981169] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5825cc57-ad66-4c0b-b892-02aba2942239 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.985026] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 911.985026] env[62552]: value = "task-1239676" [ 911.985026] env[62552]: _type = "Task" [ 911.985026] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.995229] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239676, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.011478] env[62552]: DEBUG nova.compute.manager [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 912.011718] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 912.012627] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9335ef1d-f169-4ac0-8f9d-72f01976c5c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.020389] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.020725] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2799511-6213-47ef-b30c-64f8903c42dd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.027598] env[62552]: DEBUG oslo_vmware.api [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 912.027598] env[62552]: value = "task-1239677" [ 912.027598] env[62552]: _type = "Task" [ 912.027598] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.040368] env[62552]: DEBUG oslo_vmware.api [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239677, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.100315] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239673, 'name': ReconfigVM_Task, 'duration_secs': 0.328446} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.100828] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Reconfigured VM instance instance-0000004d to attach disk [datastore2] fbbc360d-9cfd-48f1-80b2-26da2c72c002/fbbc360d-9cfd-48f1-80b2-26da2c72c002.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.101574] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7a1866e-420d-4e5c-a278-58624814d52a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.109305] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 912.109305] env[62552]: value = "task-1239678" [ 912.109305] env[62552]: _type = "Task" [ 912.109305] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.121909] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239678, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.125443] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239674, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.346309] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.346510] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.346701] env[62552]: DEBUG nova.network.neutron [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.393264] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239675, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.399084] env[62552]: DEBUG nova.objects.base [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 912.400597] env[62552]: DEBUG nova.network.neutron [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 912.476659] env[62552]: DEBUG nova.compute.manager [req-ff8613b8-df3a-4b22-9d83-4f66c174b98b req-d6f6f384-ea0e-4e73-8bd9-7478eff9ca8a service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received event network-vif-plugged-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 912.476659] env[62552]: DEBUG oslo_concurrency.lockutils [req-ff8613b8-df3a-4b22-9d83-4f66c174b98b req-d6f6f384-ea0e-4e73-8bd9-7478eff9ca8a service nova] Acquiring lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.476659] env[62552]: DEBUG oslo_concurrency.lockutils [req-ff8613b8-df3a-4b22-9d83-4f66c174b98b req-d6f6f384-ea0e-4e73-8bd9-7478eff9ca8a service nova] Lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.476836] env[62552]: DEBUG oslo_concurrency.lockutils [req-ff8613b8-df3a-4b22-9d83-4f66c174b98b req-d6f6f384-ea0e-4e73-8bd9-7478eff9ca8a service nova] Lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.477045] env[62552]: DEBUG nova.compute.manager [req-ff8613b8-df3a-4b22-9d83-4f66c174b98b req-d6f6f384-ea0e-4e73-8bd9-7478eff9ca8a service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] No waiting events found dispatching network-vif-plugged-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 912.477693] env[62552]: WARNING nova.compute.manager [req-ff8613b8-df3a-4b22-9d83-4f66c174b98b req-d6f6f384-ea0e-4e73-8bd9-7478eff9ca8a service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received unexpected event network-vif-plugged-929d55b0-5a35-4d2d-a172-15d79a215977 for instance with vm_state shelved_offloaded and task_state spawning. [ 912.496831] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239676, 'name': Rename_Task, 'duration_secs': 0.167029} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.499858] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 912.500383] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b81310dd-c491-4971-9a41-849ff13467c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.507176] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 912.507176] env[62552]: value = "task-1239679" [ 912.507176] env[62552]: _type = "Task" [ 912.507176] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.519305] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239679, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.521162] env[62552]: DEBUG nova.policy [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6515846d33e44bca942767f9e1d650b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2f57f72de0c4adc9754458f4e7040b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 912.540289] env[62552]: DEBUG oslo_vmware.api [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239677, 'name': PowerOffVM_Task, 'duration_secs': 0.202036} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.540707] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 912.540818] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 912.541198] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0709ae5-e4ff-44b3-9583-34f6362f7da4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.598958] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 912.599738] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 912.599974] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Deleting the datastore file [datastore1] 43d1e732-7e9b-4f9f-b67c-f0ef2be91902 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.600285] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0136262-b11f-4d0f-8012-59cd4a5a4844 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.608665] env[62552]: DEBUG oslo_vmware.api [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for the task: (returnval){ [ 912.608665] env[62552]: value = "task-1239681" [ 912.608665] env[62552]: _type = "Task" [ 912.608665] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.632930] env[62552]: DEBUG oslo_vmware.api [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239681, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.642778] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239678, 'name': Rename_Task, 'duration_secs': 0.186289} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.643110] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239674, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562774} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.644394] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 912.644394] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6a5b5f4e-0c84-447d-a3da-8258512abb7c/6a5b5f4e-0c84-447d-a3da-8258512abb7c.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.644394] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.644691] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40a04799-365b-46dc-b1de-09f8d0e7bdbe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.649284] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2796db48-3483-48e3-86dc-a7f0481d6e23 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.655409] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 912.655409] env[62552]: value = "task-1239683" [ 912.655409] env[62552]: _type = "Task" [ 912.655409] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.659879] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 912.659879] env[62552]: value = "task-1239682" [ 912.659879] env[62552]: _type = "Task" [ 912.659879] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.676086] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239683, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.676362] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239682, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.731039] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee75819-44ab-40b3-8041-ba4c456c1d89 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.740033] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d57090b-72f4-4b5e-ab36-c23db8bfe0d5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.780562] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e1ec0d-b15e-44db-b891-cc33d9832437 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.785407] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0827ea18-9482-4ff4-93ba-9603459a0137 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.801756] env[62552]: DEBUG nova.compute.provider_tree [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.894651] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239675, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.707363} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.895024] env[62552]: INFO nova.virt.vmwareapi.ds_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 6c8b5367-4fa0-479b-9382-ff261201e3ef/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. [ 912.897532] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9023f7f0-bbe3-4b6e-9845-14a298ef6579 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.926718] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 6c8b5367-4fa0-479b-9382-ff261201e3ef/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.927117] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cd4ea77-1eb3-4f06-8770-943aadeeca1b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.946721] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 912.946721] env[62552]: value = "task-1239684" [ 912.946721] env[62552]: _type = "Task" [ 912.946721] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.958914] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.020302] env[62552]: DEBUG oslo_vmware.api [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239679, 'name': PowerOnVM_Task, 'duration_secs': 0.510975} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.020769] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 913.021085] env[62552]: INFO nova.compute.manager [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Took 8.94 seconds to spawn the instance on the hypervisor. [ 913.021359] env[62552]: DEBUG nova.compute.manager [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 913.026021] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab40800-b2eb-48cf-ba2b-20c8c40ca63e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.120169] env[62552]: DEBUG oslo_vmware.api [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Task: {'id': task-1239681, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288036} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.120169] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.120169] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.120370] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.120571] env[62552]: INFO nova.compute.manager [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Took 1.11 seconds to destroy the instance on the hypervisor. [ 913.120937] env[62552]: DEBUG oslo.service.loopingcall [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.121364] env[62552]: DEBUG nova.compute.manager [-] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 913.121364] env[62552]: DEBUG nova.network.neutron [-] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 913.178072] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239683, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088201} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.181676] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.182093] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239682, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.182865] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54043a0d-0c19-43eb-8f0c-4e1c63ee0d6b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.205273] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 6a5b5f4e-0c84-447d-a3da-8258512abb7c/6a5b5f4e-0c84-447d-a3da-8258512abb7c.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.207801] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23a7bd11-a1a3-4ebf-89e5-437131770396 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.229953] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 913.229953] env[62552]: value = "task-1239685" [ 913.229953] env[62552]: _type = "Task" [ 913.229953] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.239930] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239685, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.305667] env[62552]: DEBUG nova.scheduler.client.report [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 913.330481] env[62552]: DEBUG nova.network.neutron [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating instance_info_cache with network_info: [{"id": "929d55b0-5a35-4d2d-a172-15d79a215977", "address": "fa:16:3e:62:4d:9b", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap929d55b0-5a", "ovs_interfaceid": "929d55b0-5a35-4d2d-a172-15d79a215977", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.458165] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239684, 'name': ReconfigVM_Task, 'duration_secs': 0.499959} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.458794] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 6c8b5367-4fa0-479b-9382-ff261201e3ef/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.460612] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03659b3f-413f-44f0-9490-9f581f7b8e7c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.507975] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8bc4c74-b68a-418c-bbed-abea0e57f2c7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.530576] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 913.530576] env[62552]: value = "task-1239686" [ 913.530576] env[62552]: _type = "Task" [ 913.530576] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.548832] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239686, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.549426] env[62552]: INFO nova.compute.manager [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Took 49.96 seconds to build instance. [ 913.628413] env[62552]: DEBUG oslo_vmware.rw_handles [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52722003-8448-e6b5-2a48-4c25bdfb3c62/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 913.629394] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce96701-33e3-4c04-998f-9de040f81c13 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.637235] env[62552]: DEBUG oslo_vmware.rw_handles [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52722003-8448-e6b5-2a48-4c25bdfb3c62/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 913.637412] env[62552]: ERROR oslo_vmware.rw_handles [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52722003-8448-e6b5-2a48-4c25bdfb3c62/disk-0.vmdk due to incomplete transfer. [ 913.637668] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0751fcdd-3e52-477a-a1b5-5d8166ace2ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.644326] env[62552]: DEBUG oslo_vmware.rw_handles [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52722003-8448-e6b5-2a48-4c25bdfb3c62/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 913.644890] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Uploaded image 9f1bf645-d022-4cd1-9ecf-2fc47981b465 to the Glance image server {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 913.647368] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 913.647763] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cb13f092-116e-46a3-bb27-1ff17ef92ff9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.653778] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 913.653778] env[62552]: value = "task-1239687" [ 913.653778] env[62552]: _type = "Task" [ 913.653778] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.663719] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239687, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.678569] env[62552]: DEBUG oslo_vmware.api [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239682, 'name': PowerOnVM_Task, 'duration_secs': 0.660649} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.678876] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 913.679459] env[62552]: INFO nova.compute.manager [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Took 12.12 seconds to spawn the instance on the hypervisor. [ 913.679459] env[62552]: DEBUG nova.compute.manager [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 913.680351] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9fb0ebb-ccc5-4298-8526-ad3f85e7620c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.740679] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239685, 'name': ReconfigVM_Task, 'duration_secs': 0.416336} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.741042] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 6a5b5f4e-0c84-447d-a3da-8258512abb7c/6a5b5f4e-0c84-447d-a3da-8258512abb7c.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.741862] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b775ff78-3c1b-4348-833a-39ba6e3f9b32 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.754740] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 913.754740] env[62552]: value = "task-1239688" [ 913.754740] env[62552]: _type = "Task" [ 913.754740] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.763397] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239688, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.812449] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.813052] env[62552]: DEBUG nova.compute.manager [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 913.815669] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.446s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.815896] env[62552]: DEBUG nova.objects.instance [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lazy-loading 'resources' on Instance uuid fb47f136-7859-4756-bc92-fbcf430b4c8b {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.835766] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.861706] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='2127c895f9f8e3c2d4314edeb97410f9',container_format='bare',created_at=2024-10-10T11:35:16Z,direct_url=,disk_format='vmdk',id=d88505cd-62b9-4da3-926e-217d15bab510,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1185004650-shelved',owner='25030025e90243b290a7d90efa26ec79',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-10T11:35:33Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.862482] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.862482] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.862482] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.862482] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.862690] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.862857] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.863171] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.863408] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.863572] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.863767] env[62552]: DEBUG nova.virt.hardware [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.865439] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ed7e6f-c6e6-4f0b-a4b0-0d92b581d02d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.874465] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228b5911-203f-4c04-9c64-3f07f676e912 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.891157] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:4d:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '929d55b0-5a35-4d2d-a172-15d79a215977', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.899471] env[62552]: DEBUG oslo.service.loopingcall [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.899471] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 913.899848] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-231ddac7-7fa2-427b-9cde-9a8fa338f594 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.923038] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.923038] env[62552]: value = "task-1239689" [ 913.923038] env[62552]: _type = "Task" [ 913.923038] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.930250] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239689, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.041552] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239686, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.042934] env[62552]: DEBUG nova.network.neutron [-] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.053263] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dfa725cd-ec2a-46d9-81d9-5276b50ce1ec tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "0080a6b3-cac0-44f3-a037-3f94f1daa275" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.488s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.171437] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239687, 'name': Destroy_Task, 'duration_secs': 0.335851} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.171437] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Destroyed the VM [ 914.171648] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 914.171973] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f9aa87d0-997b-417c-920a-a73e943376f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.178943] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 914.178943] env[62552]: value = "task-1239690" [ 914.178943] env[62552]: _type = "Task" [ 914.178943] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.189301] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239690, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.199023] env[62552]: INFO nova.compute.manager [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Took 52.20 seconds to build instance. [ 914.230486] env[62552]: DEBUG nova.network.neutron [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Successfully updated port: 9726e936-f144-4cce-9964-78bcae1c40ae {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.266373] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239688, 'name': Rename_Task, 'duration_secs': 0.163075} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.266597] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.267788] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6593d4d2-2b62-4993-8621-da47064ca943 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.274583] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 914.274583] env[62552]: value = "task-1239691" [ 914.274583] env[62552]: _type = "Task" [ 914.274583] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.283412] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239691, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.285551] env[62552]: DEBUG nova.compute.manager [req-f8cfa9d9-a5b2-41ae-abbf-234a698ef5e1 req-fbe481eb-131e-4f8f-8dfc-1d39e272d8fb service nova] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Received event network-vif-deleted-a9ece5cd-e814-4895-b895-6ac928e22cfb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 914.318869] env[62552]: DEBUG nova.compute.utils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.324127] env[62552]: DEBUG nova.compute.manager [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 914.324328] env[62552]: DEBUG nova.network.neutron [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 914.368932] env[62552]: DEBUG nova.policy [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7f8aa8200874dddb71d8b21bd12ca04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8b8d96b464a439e9c7ef6f3e419a9bc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 914.434376] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239689, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.545368] env[62552]: INFO nova.compute.manager [-] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Took 1.42 seconds to deallocate network for instance. [ 914.545715] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239686, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.659147] env[62552]: DEBUG nova.network.neutron [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Successfully created port: 61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.689607] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239690, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.701638] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3d7bb150-e37d-448b-a96d-140a5dbf4cd8 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.434s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.732376] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.732580] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.732780] env[62552]: DEBUG nova.network.neutron [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.768215] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f6e217-21c5-4893-a37a-16e151bc45b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.789300] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23ac9e6-1d9c-4bcf-901e-d95a684cc14d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.800185] env[62552]: DEBUG oslo_vmware.api [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239691, 'name': PowerOnVM_Task, 'duration_secs': 0.500413} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.827330] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.827330] env[62552]: INFO nova.compute.manager [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Took 8.20 seconds to spawn the instance on the hypervisor. [ 914.827330] env[62552]: DEBUG nova.compute.manager [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 914.828743] env[62552]: DEBUG nova.compute.manager [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 914.835022] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154609b3-5aa7-43a2-bf2e-8233946e6921 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.835485] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad58b4c5-b13b-444d-a337-d4c591ffbe8f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.848616] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9707a58-37c6-45e5-9bb9-b7253fa9e3b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.865249] env[62552]: DEBUG nova.compute.provider_tree [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.932421] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239689, 'name': CreateVM_Task, 'duration_secs': 0.622892} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.932686] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.933319] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.933492] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.933882] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.934150] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18e7dfcb-2bca-4d9a-8b81-d6a9630dd832 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.939321] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 914.939321] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5299813f-7838-ed5a-d884-55958ce8ab45" [ 914.939321] env[62552]: _type = "Task" [ 914.939321] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.947046] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5299813f-7838-ed5a-d884-55958ce8ab45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.042439] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239686, 'name': ReconfigVM_Task, 'duration_secs': 1.16892} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.042715] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.042968] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb7769ac-6c79-4239-b9dd-eecfc2cbc158 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.049963] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 915.049963] env[62552]: value = "task-1239692" [ 915.049963] env[62552]: _type = "Task" [ 915.049963] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.056528] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.061540] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.190409] env[62552]: DEBUG oslo_vmware.api [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239690, 'name': RemoveSnapshot_Task, 'duration_secs': 0.723806} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.190594] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 915.190678] env[62552]: INFO nova.compute.manager [None req-3900c0fc-eb42-4f8a-bc4c-04138fb6dac0 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Took 16.84 seconds to snapshot the instance on the hypervisor. [ 915.282297] env[62552]: WARNING nova.network.neutron [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] 06f1ea1a-5103-419f-a356-3b65b9a2685e already exists in list: networks containing: ['06f1ea1a-5103-419f-a356-3b65b9a2685e']. ignoring it [ 915.282956] env[62552]: WARNING nova.network.neutron [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] 06f1ea1a-5103-419f-a356-3b65b9a2685e already exists in list: networks containing: ['06f1ea1a-5103-419f-a356-3b65b9a2685e']. ignoring it [ 915.354951] env[62552]: INFO nova.compute.manager [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Took 43.39 seconds to build instance. [ 915.367599] env[62552]: DEBUG nova.scheduler.client.report [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 915.452949] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.453231] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Processing image d88505cd-62b9-4da3-926e-217d15bab510 {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.453465] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510/d88505cd-62b9-4da3-926e-217d15bab510.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.453615] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510/d88505cd-62b9-4da3-926e-217d15bab510.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.453790] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.454073] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1bb96e50-ccf5-4690-9f3d-474544f2d95e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.471032] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.471270] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.472980] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b3a8b8e-1fa7-4ac6-a891-3f4d492decef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.477200] env[62552]: DEBUG nova.compute.manager [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received event network-changed-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 915.477386] env[62552]: DEBUG nova.compute.manager [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Refreshing instance network info cache due to event network-changed-929d55b0-5a35-4d2d-a172-15d79a215977. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 915.477601] env[62552]: DEBUG oslo_concurrency.lockutils [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] Acquiring lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.477818] env[62552]: DEBUG oslo_concurrency.lockutils [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] Acquired lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.477907] env[62552]: DEBUG nova.network.neutron [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Refreshing network info cache for port 929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.482028] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 915.482028] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c69bb1-bfac-7966-42dc-b7661c244294" [ 915.482028] env[62552]: _type = "Task" [ 915.482028] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.492443] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c69bb1-bfac-7966-42dc-b7661c244294, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.560009] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239692, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.839973] env[62552]: DEBUG nova.compute.manager [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 915.857731] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4f039833-ab03-432b-846a-806c06a715de tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.908s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.866885] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.867178] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.867351] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.867541] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.867692] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.867844] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.868420] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.868668] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.869049] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.869381] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.869594] env[62552]: DEBUG nova.virt.hardware [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.871052] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bc6d29-f33f-4aa0-a9c5-0dd83a9c1d31 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.875538] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.060s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.878199] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.101s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.878440] env[62552]: DEBUG nova.objects.instance [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lazy-loading 'resources' on Instance uuid fea83dde-3181-49dd-a000-5e5cffc8de95 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.890090] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfd6feb-fc55-4dd1-8ea8-1cacb0285025 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.906371] env[62552]: INFO nova.scheduler.client.report [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Deleted allocations for instance fb47f136-7859-4756-bc92-fbcf430b4c8b [ 915.942062] env[62552]: DEBUG nova.network.neutron [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [{"id": "0679d99d-6c79-408f-82a1-9f80978a1c48", "address": "fa:16:3e:c1:12:aa", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0679d99d-6c", "ovs_interfaceid": "0679d99d-6c79-408f-82a1-9f80978a1c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "035ebcb4-aada-41ce-9361-47d4162965cf", "address": "fa:16:3e:5c:87:4a", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap035ebcb4-aa", "ovs_interfaceid": "035ebcb4-aada-41ce-9361-47d4162965cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9726e936-f144-4cce-9964-78bcae1c40ae", "address": "fa:16:3e:28:e5:36", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9726e936-f1", "ovs_interfaceid": "9726e936-f144-4cce-9964-78bcae1c40ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.993016] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Preparing fetch location {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 915.993628] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Fetch image to [datastore1] OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba/OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba.vmdk {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 915.994178] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Downloading stream optimized image d88505cd-62b9-4da3-926e-217d15bab510 to [datastore1] OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba/OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba.vmdk on the data store datastore1 as vApp {{(pid=62552) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 915.996365] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Downloading image file data d88505cd-62b9-4da3-926e-217d15bab510 to the ESX as VM named 'OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba' {{(pid=62552) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 916.062721] env[62552]: DEBUG oslo_vmware.api [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239692, 'name': PowerOnVM_Task, 'duration_secs': 0.705147} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.062721] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.067757] env[62552]: DEBUG nova.compute.manager [None req-70fb894a-57b3-4f0b-bb2d-15a5ad7d4415 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 916.070209] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-707750b4-76da-4794-8adb-e884aab847ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.094016] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 916.094016] env[62552]: value = "resgroup-9" [ 916.094016] env[62552]: _type = "ResourcePool" [ 916.094016] env[62552]: }. {{(pid=62552) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 916.094016] env[62552]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-13406c4c-d81c-4a52-bc6e-87cdadbd9e1f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.137743] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lease: (returnval){ [ 916.137743] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52659f5f-5029-cb53-dab1-3357366bf6d9" [ 916.137743] env[62552]: _type = "HttpNfcLease" [ 916.137743] env[62552]: } obtained for vApp import into resource pool (val){ [ 916.137743] env[62552]: value = "resgroup-9" [ 916.137743] env[62552]: _type = "ResourcePool" [ 916.137743] env[62552]: }. {{(pid=62552) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 916.137743] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the lease: (returnval){ [ 916.137743] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52659f5f-5029-cb53-dab1-3357366bf6d9" [ 916.137743] env[62552]: _type = "HttpNfcLease" [ 916.137743] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 916.147370] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.147370] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52659f5f-5029-cb53-dab1-3357366bf6d9" [ 916.147370] env[62552]: _type = "HttpNfcLease" [ 916.147370] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 916.214434] env[62552]: DEBUG nova.network.neutron [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Successfully updated port: 61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 916.322756] env[62552]: DEBUG nova.network.neutron [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updated VIF entry in instance network info cache for port 929d55b0-5a35-4d2d-a172-15d79a215977. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.323271] env[62552]: DEBUG nova.network.neutron [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating instance_info_cache with network_info: [{"id": "929d55b0-5a35-4d2d-a172-15d79a215977", "address": "fa:16:3e:62:4d:9b", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap929d55b0-5a", "ovs_interfaceid": "929d55b0-5a35-4d2d-a172-15d79a215977", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.419027] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a9200c7b-3561-4e4f-aa2a-f5586951fb3f tempest-ServerAddressesTestJSON-266115812 tempest-ServerAddressesTestJSON-266115812-project-member] Lock "fb47f136-7859-4756-bc92-fbcf430b4c8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.610s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.444729] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.445479] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.445665] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.446792] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1058d3cf-627b-4619-af8d-26c12a8a1c71 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.467320] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.467631] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.467803] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.468051] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.468232] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.468425] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.468698] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.468889] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.469109] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.469368] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.469583] env[62552]: DEBUG nova.virt.hardware [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.475940] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Reconfiguring VM to attach interface {{(pid=62552) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 916.476324] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-571a2db0-9cdc-4607-94ed-c249c019cad5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.497100] env[62552]: DEBUG oslo_vmware.api [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 916.497100] env[62552]: value = "task-1239694" [ 916.497100] env[62552]: _type = "Task" [ 916.497100] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.508401] env[62552]: DEBUG oslo_vmware.api [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239694, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.651167] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.651167] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52659f5f-5029-cb53-dab1-3357366bf6d9" [ 916.651167] env[62552]: _type = "HttpNfcLease" [ 916.651167] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 916.721517] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.721715] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.726762] env[62552]: DEBUG nova.network.neutron [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.741245] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d645bd-d9ed-4236-a7a7-3a47b165a508 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.751507] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6f0a96-7b6b-41b7-8b88-1e0b9130ae8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.789271] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61f745f-5b33-4fcf-8170-d300eb09b581 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.797549] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ff0d3b-55c3-4860-8612-ca69192d5556 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.812600] env[62552]: DEBUG nova.compute.provider_tree [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.825954] env[62552]: DEBUG oslo_concurrency.lockutils [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] Releasing lock "refresh_cache-cdee22a9-4327-47af-ab14-8403a06b2802" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.826249] env[62552]: DEBUG nova.compute.manager [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-vif-plugged-9726e936-f144-4cce-9964-78bcae1c40ae {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 916.826453] env[62552]: DEBUG oslo_concurrency.lockutils [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.826661] env[62552]: DEBUG oslo_concurrency.lockutils [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.826828] env[62552]: DEBUG oslo_concurrency.lockutils [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.826997] env[62552]: DEBUG nova.compute.manager [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] No waiting events found dispatching network-vif-plugged-9726e936-f144-4cce-9964-78bcae1c40ae {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 916.827196] env[62552]: WARNING nova.compute.manager [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received unexpected event network-vif-plugged-9726e936-f144-4cce-9964-78bcae1c40ae for instance with vm_state active and task_state None. [ 916.827359] env[62552]: DEBUG nova.compute.manager [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-changed-9726e936-f144-4cce-9964-78bcae1c40ae {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 916.827520] env[62552]: DEBUG nova.compute.manager [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Refreshing instance network info cache due to event network-changed-9726e936-f144-4cce-9964-78bcae1c40ae. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 916.827707] env[62552]: DEBUG oslo_concurrency.lockutils [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] Acquiring lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.827883] env[62552]: DEBUG oslo_concurrency.lockutils [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] Acquired lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.828016] env[62552]: DEBUG nova.network.neutron [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Refreshing network info cache for port 9726e936-f144-4cce-9964-78bcae1c40ae {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.008374] env[62552]: DEBUG oslo_vmware.api [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239694, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.158746] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 917.158746] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52659f5f-5029-cb53-dab1-3357366bf6d9" [ 917.158746] env[62552]: _type = "HttpNfcLease" [ 917.158746] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 917.159116] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 917.159116] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52659f5f-5029-cb53-dab1-3357366bf6d9" [ 917.159116] env[62552]: _type = "HttpNfcLease" [ 917.159116] env[62552]: }. {{(pid=62552) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 917.160208] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44699415-48ba-4890-ad07-df4b3442df2e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.171310] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a4d8ad-d079-61f5-2a30-c57c1f836f7d/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 917.171591] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a4d8ad-d079-61f5-2a30-c57c1f836f7d/disk-0.vmdk. {{(pid=62552) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 917.246437] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5113db20-ad3b-42a9-a3a7-d6f748a947ee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.286125] env[62552]: DEBUG nova.network.neutron [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.316481] env[62552]: DEBUG nova.scheduler.client.report [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 917.518627] env[62552]: DEBUG oslo_vmware.api [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239694, 'name': ReconfigVM_Task, 'duration_secs': 0.785243} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.521298] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.521560] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Reconfigured VM to attach interface {{(pid=62552) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 917.574101] env[62552]: DEBUG nova.network.neutron [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Updating instance_info_cache with network_info: [{"id": "61514a33-d741-4aae-876d-b0ecbe6ed6e6", "address": "fa:16:3e:b9:cf:bd", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61514a33-d7", "ovs_interfaceid": "61514a33-d741-4aae-876d-b0ecbe6ed6e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.737945] env[62552]: DEBUG nova.network.neutron [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updated VIF entry in instance network info cache for port 9726e936-f144-4cce-9964-78bcae1c40ae. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 917.738555] env[62552]: DEBUG nova.network.neutron [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [{"id": "0679d99d-6c79-408f-82a1-9f80978a1c48", "address": "fa:16:3e:c1:12:aa", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0679d99d-6c", "ovs_interfaceid": "0679d99d-6c79-408f-82a1-9f80978a1c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "035ebcb4-aada-41ce-9361-47d4162965cf", "address": "fa:16:3e:5c:87:4a", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap035ebcb4-aa", "ovs_interfaceid": "035ebcb4-aada-41ce-9361-47d4162965cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9726e936-f144-4cce-9964-78bcae1c40ae", "address": "fa:16:3e:28:e5:36", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9726e936-f1", "ovs_interfaceid": "9726e936-f144-4cce-9964-78bcae1c40ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.822777] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.944s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.825513] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.315s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.829476] env[62552]: INFO nova.compute.claims [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.864755] env[62552]: INFO nova.scheduler.client.report [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Deleted allocations for instance fea83dde-3181-49dd-a000-5e5cffc8de95 [ 918.030160] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d0eafb9-4f2d-4335-bcbd-5f74cf12d42b tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-9726e936-f144-4cce-9964-78bcae1c40ae" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.809s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.076149] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.076149] env[62552]: DEBUG nova.compute.manager [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Instance network_info: |[{"id": "61514a33-d741-4aae-876d-b0ecbe6ed6e6", "address": "fa:16:3e:b9:cf:bd", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61514a33-d7", "ovs_interfaceid": "61514a33-d741-4aae-876d-b0ecbe6ed6e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 918.078208] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:cf:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61514a33-d741-4aae-876d-b0ecbe6ed6e6', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.087456] env[62552]: DEBUG oslo.service.loopingcall [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.090414] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.091817] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6aac85e4-b8e9-46ec-a02f-8a119a897fd6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.115333] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.115333] env[62552]: value = "task-1239695" [ 918.115333] env[62552]: _type = "Task" [ 918.115333] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.125494] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239695, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.231301] env[62552]: DEBUG nova.compute.manager [req-e0e6f940-3b75-4ad0-9266-2d49a48502d4 req-c9db9a73-5019-458c-b9b1-04b082546bfe service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Received event network-changed-5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 918.231301] env[62552]: DEBUG nova.compute.manager [req-e0e6f940-3b75-4ad0-9266-2d49a48502d4 req-c9db9a73-5019-458c-b9b1-04b082546bfe service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Refreshing instance network info cache due to event network-changed-5e69a23a-a63b-4d24-9729-859b537f945b. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 918.231301] env[62552]: DEBUG oslo_concurrency.lockutils [req-e0e6f940-3b75-4ad0-9266-2d49a48502d4 req-c9db9a73-5019-458c-b9b1-04b082546bfe service nova] Acquiring lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.231405] env[62552]: DEBUG oslo_concurrency.lockutils [req-e0e6f940-3b75-4ad0-9266-2d49a48502d4 req-c9db9a73-5019-458c-b9b1-04b082546bfe service nova] Acquired lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.231505] env[62552]: DEBUG nova.network.neutron [req-e0e6f940-3b75-4ad0-9266-2d49a48502d4 req-c9db9a73-5019-458c-b9b1-04b082546bfe service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Refreshing network info cache for port 5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.242898] env[62552]: DEBUG oslo_concurrency.lockutils [req-4c6f5f6c-9a8b-4c84-a6f6-e8376d43791c req-333ae17d-b6d7-4452-b7be-2cd24bc9fbf9 service nova] Releasing lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.328239] env[62552]: DEBUG nova.compute.manager [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Received event network-vif-plugged-61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 918.328597] env[62552]: DEBUG oslo_concurrency.lockutils [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] Acquiring lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.329246] env[62552]: DEBUG oslo_concurrency.lockutils [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.329246] env[62552]: DEBUG oslo_concurrency.lockutils [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.329387] env[62552]: DEBUG nova.compute.manager [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] No waiting events found dispatching network-vif-plugged-61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 918.329659] env[62552]: WARNING nova.compute.manager [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Received unexpected event network-vif-plugged-61514a33-d741-4aae-876d-b0ecbe6ed6e6 for instance with vm_state building and task_state spawning. [ 918.329872] env[62552]: DEBUG nova.compute.manager [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Received event network-changed-61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 918.330112] env[62552]: DEBUG nova.compute.manager [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Refreshing instance network info cache due to event network-changed-61514a33-d741-4aae-876d-b0ecbe6ed6e6. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 918.330347] env[62552]: DEBUG oslo_concurrency.lockutils [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] Acquiring lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.330557] env[62552]: DEBUG oslo_concurrency.lockutils [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] Acquired lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.330799] env[62552]: DEBUG nova.network.neutron [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Refreshing network info cache for port 61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.377230] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b38e094-d889-41b8-b90f-cff95cdc8649 tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fea83dde-3181-49dd-a000-5e5cffc8de95" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.132s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.589341] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Completed reading data from the image iterator. {{(pid=62552) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 918.589653] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a4d8ad-d079-61f5-2a30-c57c1f836f7d/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 918.590811] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3da198-ce95-4623-b56a-ba38a019b310 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.599865] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a4d8ad-d079-61f5-2a30-c57c1f836f7d/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 918.600530] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a4d8ad-d079-61f5-2a30-c57c1f836f7d/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 918.600898] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b9e30185-e831-4dec-8880-7e33152d2fd4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.626035] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239695, 'name': CreateVM_Task, 'duration_secs': 0.4314} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.626320] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.627146] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.627393] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.627801] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.628132] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ff87c4b-4c43-4018-9d34-7b704a757ef8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.633054] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 918.633054] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5272340b-bc33-66da-edce-d31201d3efb2" [ 918.633054] env[62552]: _type = "Task" [ 918.633054] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.642470] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5272340b-bc33-66da-edce-d31201d3efb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.998926] env[62552]: DEBUG oslo_vmware.rw_handles [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a4d8ad-d079-61f5-2a30-c57c1f836f7d/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 918.999383] env[62552]: INFO nova.virt.vmwareapi.images [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Downloaded image file data d88505cd-62b9-4da3-926e-217d15bab510 [ 919.001799] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d15487-733f-476e-bb3c-698825c87c4c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.027747] env[62552]: DEBUG nova.network.neutron [req-e0e6f940-3b75-4ad0-9266-2d49a48502d4 req-c9db9a73-5019-458c-b9b1-04b082546bfe service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updated VIF entry in instance network info cache for port 5e69a23a-a63b-4d24-9729-859b537f945b. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 919.028127] env[62552]: DEBUG nova.network.neutron [req-e0e6f940-3b75-4ad0-9266-2d49a48502d4 req-c9db9a73-5019-458c-b9b1-04b082546bfe service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updating instance_info_cache with network_info: [{"id": "5e69a23a-a63b-4d24-9729-859b537f945b", "address": "fa:16:3e:0b:c1:98", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e69a23a-a6", "ovs_interfaceid": "5e69a23a-a63b-4d24-9729-859b537f945b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.029430] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ca0a51b-18d7-4da4-bee1-08a1d5cedfb8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.114051] env[62552]: INFO nova.virt.vmwareapi.images [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] The imported VM was unregistered [ 919.116146] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Caching image {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 919.117723] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating directory with path [datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510 {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.119616] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26703218-cde0-4034-842a-19175730a31e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.144635] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5272340b-bc33-66da-edce-d31201d3efb2, 'name': SearchDatastore_Task, 'duration_secs': 0.019942} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.147220] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.147489] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.147769] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.147925] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.148537] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.151360] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5249c6d4-cbc9-43b2-8547-2af052925100 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.153284] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Created directory with path [datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510 {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.153474] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba/OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba.vmdk to [datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510/d88505cd-62b9-4da3-926e-217d15bab510.vmdk. {{(pid=62552) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 919.153718] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ba761538-bd95-4073-9dc6-58e8a24d5975 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.167096] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 919.167096] env[62552]: value = "task-1239697" [ 919.167096] env[62552]: _type = "Task" [ 919.167096] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.171357] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.171594] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.172760] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bfc5b6a-e03a-496d-83d1-1fd96120dadf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.180698] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239697, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.186682] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 919.186682] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5270abef-1b3b-7bb9-bb94-23168e684bff" [ 919.186682] env[62552]: _type = "Task" [ 919.186682] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.194754] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5270abef-1b3b-7bb9-bb94-23168e684bff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.256649] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027d9df0-e86f-47e9-97bf-3e04135ad863 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.265386] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703af97d-64b1-48f3-bc1d-902d4381329a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.299214] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fe5457-9d88-41b2-a744-9faf717dd989 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.302186] env[62552]: DEBUG nova.network.neutron [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Updated VIF entry in instance network info cache for port 61514a33-d741-4aae-876d-b0ecbe6ed6e6. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 919.302624] env[62552]: DEBUG nova.network.neutron [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Updating instance_info_cache with network_info: [{"id": "61514a33-d741-4aae-876d-b0ecbe6ed6e6", "address": "fa:16:3e:b9:cf:bd", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61514a33-d7", "ovs_interfaceid": "61514a33-d741-4aae-876d-b0ecbe6ed6e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.312223] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fec832b-2696-447b-ba5b-c68a78a1ef86 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.329672] env[62552]: DEBUG nova.compute.provider_tree [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.511894] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "fb881c34-dccc-4703-af8e-c75caafd9b08" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.512818] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fb881c34-dccc-4703-af8e-c75caafd9b08" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.512818] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "fb881c34-dccc-4703-af8e-c75caafd9b08-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.512818] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fb881c34-dccc-4703-af8e-c75caafd9b08-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.512818] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fb881c34-dccc-4703-af8e-c75caafd9b08-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.518887] env[62552]: INFO nova.compute.manager [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Terminating instance [ 919.532576] env[62552]: DEBUG oslo_concurrency.lockutils [req-e0e6f940-3b75-4ad0-9266-2d49a48502d4 req-c9db9a73-5019-458c-b9b1-04b082546bfe service nova] Releasing lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.676688] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239697, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.697938] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5270abef-1b3b-7bb9-bb94-23168e684bff, 'name': SearchDatastore_Task, 'duration_secs': 0.048839} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.698706] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65cd79a4-105a-4318-9818-834c9afc1950 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.703892] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 919.703892] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520cb165-37d9-391c-4d90-c19e004b0a82" [ 919.703892] env[62552]: _type = "Task" [ 919.703892] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.711229] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520cb165-37d9-391c-4d90-c19e004b0a82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.805736] env[62552]: DEBUG oslo_concurrency.lockutils [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] Releasing lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.806071] env[62552]: DEBUG nova.compute.manager [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Received event network-changed-9136fa4c-29b1-49d3-ba65-6530ce9441b8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 919.806261] env[62552]: DEBUG nova.compute.manager [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Refreshing instance network info cache due to event network-changed-9136fa4c-29b1-49d3-ba65-6530ce9441b8. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 919.806484] env[62552]: DEBUG oslo_concurrency.lockutils [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] Acquiring lock "refresh_cache-fbbc360d-9cfd-48f1-80b2-26da2c72c002" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.806632] env[62552]: DEBUG oslo_concurrency.lockutils [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] Acquired lock "refresh_cache-fbbc360d-9cfd-48f1-80b2-26da2c72c002" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.806797] env[62552]: DEBUG nova.network.neutron [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Refreshing network info cache for port 9136fa4c-29b1-49d3-ba65-6530ce9441b8 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.832877] env[62552]: DEBUG nova.scheduler.client.report [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 920.023851] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "refresh_cache-fb881c34-dccc-4703-af8e-c75caafd9b08" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.024103] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquired lock "refresh_cache-fb881c34-dccc-4703-af8e-c75caafd9b08" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.024310] env[62552]: DEBUG nova.network.neutron [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.179248] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239697, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.219193] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520cb165-37d9-391c-4d90-c19e004b0a82, 'name': SearchDatastore_Task, 'duration_secs': 0.024765} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.219506] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.219832] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] e8e8dec1-4210-44d2-a3f8-76055f2c3b57/e8e8dec1-4210-44d2-a3f8-76055f2c3b57.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.220182] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4888853-2ceb-42f1-8046-c8bfb7fadb7a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.232655] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 920.232655] env[62552]: value = "task-1239698" [ 920.232655] env[62552]: _type = "Task" [ 920.232655] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.243711] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.323514] env[62552]: DEBUG oslo_concurrency.lockutils [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-035ebcb4-aada-41ce-9361-47d4162965cf" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.323514] env[62552]: DEBUG oslo_concurrency.lockutils [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-035ebcb4-aada-41ce-9361-47d4162965cf" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.337879] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.338462] env[62552]: DEBUG nova.compute.manager [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 920.342882] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.991s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.342974] env[62552]: DEBUG nova.objects.instance [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lazy-loading 'resources' on Instance uuid 526d0f25-3e99-4558-94c0-754ec2a80bad {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.546174] env[62552]: DEBUG nova.network.neutron [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 920.613987] env[62552]: DEBUG nova.network.neutron [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.673519] env[62552]: DEBUG nova.network.neutron [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Updated VIF entry in instance network info cache for port 9136fa4c-29b1-49d3-ba65-6530ce9441b8. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 920.673888] env[62552]: DEBUG nova.network.neutron [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Updating instance_info_cache with network_info: [{"id": "9136fa4c-29b1-49d3-ba65-6530ce9441b8", "address": "fa:16:3e:bf:29:cb", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9136fa4c-29", "ovs_interfaceid": "9136fa4c-29b1-49d3-ba65-6530ce9441b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.683112] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239697, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.747230] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.827026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.827293] env[62552]: DEBUG oslo_concurrency.lockutils [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.828176] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71157e4-5beb-4c7a-b57e-dc35747357c1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.849839] env[62552]: DEBUG nova.compute.utils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.854755] env[62552]: DEBUG nova.compute.manager [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 920.855161] env[62552]: DEBUG nova.network.neutron [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 920.857634] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035e63c5-de57-4df4-8679-61d76e9166cb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.894388] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Reconfiguring VM to detach interface {{(pid=62552) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 920.898484] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4d833e5-0514-4acc-ace8-c248ad2f6e81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.922457] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 920.922457] env[62552]: value = "task-1239699" [ 920.922457] env[62552]: _type = "Task" [ 920.922457] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.935859] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.943740] env[62552]: DEBUG nova.policy [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3faa654bd214bbb886d692e9ea29bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'adb0b1256dd34e33913d5ae5bb3553b5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.991725] env[62552]: INFO nova.compute.manager [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Rescuing [ 920.992720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.992720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.992720] env[62552]: DEBUG nova.network.neutron [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 921.119725] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Releasing lock "refresh_cache-fb881c34-dccc-4703-af8e-c75caafd9b08" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.119982] env[62552]: DEBUG nova.compute.manager [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 921.120255] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 921.121517] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194358c7-1caf-4756-9c26-53acc038f646 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.131326] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.135629] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-893941a3-59d1-4ef0-a46c-08427e63e553 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.146526] env[62552]: DEBUG oslo_vmware.api [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 921.146526] env[62552]: value = "task-1239700" [ 921.146526] env[62552]: _type = "Task" [ 921.146526] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.161576] env[62552]: DEBUG oslo_vmware.api [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239700, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.177476] env[62552]: DEBUG oslo_concurrency.lockutils [req-c7aa22be-0b86-4e9c-a940-c19d00ee171d req-cfedefaf-6f20-4099-93d8-875198e2aeed service nova] Releasing lock "refresh_cache-fbbc360d-9cfd-48f1-80b2-26da2c72c002" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.181509] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239697, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.252454] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.286335] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf64461-7ee9-490d-b7bb-7ef806074600 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.297115] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d58fb0-5725-4bec-b955-e8037ce7a6a9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.335464] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe69ddf-47c4-4d1d-9303-e877de4e5d70 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.345511] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ae1a75-783f-49bd-a8e5-63e5090d88aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.363610] env[62552]: DEBUG nova.compute.manager [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 921.367273] env[62552]: DEBUG nova.compute.provider_tree [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.436679] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.522652] env[62552]: DEBUG nova.network.neutron [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Successfully created port: cd1b0e6a-5546-47af-84df-a5e05587638d {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.661909] env[62552]: DEBUG oslo_vmware.api [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239700, 'name': PowerOffVM_Task, 'duration_secs': 0.167628} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.662264] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.662480] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.662749] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fac539c5-289d-4db3-8069-0bc40842bc08 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.685138] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239697, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.691909] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.692463] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.692463] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Deleting the datastore file [datastore2] fb881c34-dccc-4703-af8e-c75caafd9b08 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.692633] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f11b609-b72c-4882-8cc5-689522cf02a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.703831] env[62552]: DEBUG oslo_vmware.api [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for the task: (returnval){ [ 921.703831] env[62552]: value = "task-1239702" [ 921.703831] env[62552]: _type = "Task" [ 921.703831] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.714714] env[62552]: DEBUG oslo_vmware.api [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.732722] env[62552]: DEBUG nova.network.neutron [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Updating instance_info_cache with network_info: [{"id": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "address": "fa:16:3e:ce:cc:01", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap123341c8-40", "ovs_interfaceid": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.758429] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239698, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.873132] env[62552]: DEBUG nova.scheduler.client.report [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 921.940199] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.067085] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.067401] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.067636] env[62552]: INFO nova.compute.manager [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Rebooting instance [ 922.179682] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239697, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.216084] env[62552]: DEBUG oslo_vmware.api [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Task: {'id': task-1239702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26194} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.216605] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.216605] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.216728] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.216920] env[62552]: INFO nova.compute.manager [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Took 1.10 seconds to destroy the instance on the hypervisor. [ 922.217209] env[62552]: DEBUG oslo.service.loopingcall [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.217501] env[62552]: DEBUG nova.compute.manager [-] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 922.217595] env[62552]: DEBUG nova.network.neutron [-] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 922.234977] env[62552]: DEBUG nova.network.neutron [-] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 922.237313] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.252027] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239698, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.254968] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquiring lock "0080a6b3-cac0-44f3-a037-3f94f1daa275" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.254968] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "0080a6b3-cac0-44f3-a037-3f94f1daa275" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.254968] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquiring lock "0080a6b3-cac0-44f3-a037-3f94f1daa275-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.254968] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "0080a6b3-cac0-44f3-a037-3f94f1daa275-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.254968] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "0080a6b3-cac0-44f3-a037-3f94f1daa275-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.257287] env[62552]: INFO nova.compute.manager [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Terminating instance [ 922.374431] env[62552]: DEBUG nova.compute.manager [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 922.378715] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.036s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.381125] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.272s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.401418] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.401749] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.401986] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.402262] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.402463] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.402674] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.402995] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.403342] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.403573] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.403799] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.404051] env[62552]: DEBUG nova.virt.hardware [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.405211] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd35f01-50df-43a8-a4e1-54566458a731 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.411325] env[62552]: INFO nova.scheduler.client.report [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleted allocations for instance 526d0f25-3e99-4558-94c0-754ec2a80bad [ 922.421230] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ef9877-b911-4e18-9eae-f1a4a0a96aa9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.446644] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.592803] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.593074] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.593306] env[62552]: DEBUG nova.network.neutron [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.683076] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239697, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.339363} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.683257] env[62552]: INFO nova.virt.vmwareapi.ds_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba/OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba.vmdk to [datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510/d88505cd-62b9-4da3-926e-217d15bab510.vmdk. [ 922.683460] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Cleaning up location [datastore1] OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 922.683629] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_a9419afc-92b9-4988-9a95-4296c7b850ba {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.683909] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-661343bd-7312-47b1-90dd-41517edb1f2b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.690271] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 922.690271] env[62552]: value = "task-1239703" [ 922.690271] env[62552]: _type = "Task" [ 922.690271] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.698691] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239703, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.743357] env[62552]: DEBUG nova.network.neutron [-] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.750309] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239698, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.291003} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.750622] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] e8e8dec1-4210-44d2-a3f8-76055f2c3b57/e8e8dec1-4210-44d2-a3f8-76055f2c3b57.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.750853] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.751118] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92a2e2d1-8215-46d0-a07a-57c882ff6c26 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.759422] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 922.759422] env[62552]: value = "task-1239704" [ 922.759422] env[62552]: _type = "Task" [ 922.759422] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.763907] env[62552]: DEBUG nova.compute.manager [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 922.764214] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.765090] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d400c3-a3bf-4691-8d29-a1f200863332 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.773268] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.775601] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.775871] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7425c7ad-5033-446d-af03-f2bebc9e1d0c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.782195] env[62552]: DEBUG oslo_vmware.api [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 922.782195] env[62552]: value = "task-1239705" [ 922.782195] env[62552]: _type = "Task" [ 922.782195] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.794964] env[62552]: DEBUG oslo_vmware.api [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239705, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.923287] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82836846-8fbf-4779-92af-deaf68cb02ef tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "526d0f25-3e99-4558-94c0-754ec2a80bad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.072s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.937364] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.201647] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239703, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.085073} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.201980] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.202079] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510/d88505cd-62b9-4da3-926e-217d15bab510.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.202346] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510/d88505cd-62b9-4da3-926e-217d15bab510.vmdk to [datastore1] cdee22a9-4327-47af-ab14-8403a06b2802/cdee22a9-4327-47af-ab14-8403a06b2802.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.202614] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8be54d45-b65d-42ee-87d1-09ad346b1a6c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.209494] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 923.209494] env[62552]: value = "task-1239706" [ 923.209494] env[62552]: _type = "Task" [ 923.209494] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.217122] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.247731] env[62552]: INFO nova.compute.manager [-] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Took 1.03 seconds to deallocate network for instance. [ 923.268261] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.292130] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.292476] env[62552]: DEBUG oslo_vmware.api [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239705, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.292707] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1fd2851c-85b5-4de1-a0e2-7eae66a9d9e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.299341] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 923.299341] env[62552]: value = "task-1239707" [ 923.299341] env[62552]: _type = "Task" [ 923.299341] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.314346] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239707, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.344466] env[62552]: DEBUG nova.network.neutron [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.397348] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Applying migration context for instance 22357d4e-9771-477c-9fc3-fe3d76f6e902 as it has an incoming, in-progress migration ac053045-ef61-4508-bfae-a9c3df4952ed. Migration status is confirming {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 923.400511] env[62552]: INFO nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating resource usage from migration ac053045-ef61-4508-bfae-a9c3df4952ed [ 923.409702] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "4083ee43-ecea-4ea5-8923-42b348893824" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.409976] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "4083ee43-ecea-4ea5-8923-42b348893824" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.410186] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "4083ee43-ecea-4ea5-8923-42b348893824-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.410431] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "4083ee43-ecea-4ea5-8923-42b348893824-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.410544] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "4083ee43-ecea-4ea5-8923-42b348893824-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.415416] env[62552]: INFO nova.compute.manager [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Terminating instance [ 923.436501] env[62552]: DEBUG nova.network.neutron [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Successfully updated port: cd1b0e6a-5546-47af-84df-a5e05587638d {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.440980] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 0418260a-aa27-4955-ab15-b180ec04f0b3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.440980] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 714fb65d-9f80-4a81-a637-3e4398405d9b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.440980] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 34736dd0-e617-475e-baa2-cb372db1afb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 80478878-ff82-4ed6-a851-8eb2bec01e22 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 88d0f502-fc3f-429e-bdf6-a1ebacec2117 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance d8891025-5bdf-4dc9-a2b0-c86f94582ac6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 4083ee43-ecea-4ea5-8923-42b348893824 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance e31e3077-04e1-4adb-a0cc-44cf84a89eda actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance fb881c34-dccc-4703-af8e-c75caafd9b08 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: WARNING nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 5cee4242-49ae-4cb7-a208-e2982f52fbad is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 11831580-1b58-476a-91ce-a4e55947fd91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6c8b5367-4fa0-479b-9382-ff261201e3ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: WARNING nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 43d1e732-7e9b-4f9f-b67c-f0ef2be91902 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance fbbc360d-9cfd-48f1-80b2-26da2c72c002 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance cdee22a9-4327-47af-ab14-8403a06b2802 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446019] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 0080a6b3-cac0-44f3-a037-3f94f1daa275 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446668] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Migration ac053045-ef61-4508-bfae-a9c3df4952ed is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 923.446668] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 22357d4e-9771-477c-9fc3-fe3d76f6e902 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446668] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6a5b5f4e-0c84-447d-a3da-8258512abb7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446668] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance e8e8dec1-4210-44d2-a3f8-76055f2c3b57 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446668] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance ae517e5f-0cd1-457c-99a6-6cd46fdd6b06 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.446668] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 20 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 923.446668] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4416MB phys_disk=200GB used_disk=20GB total_vcpus=48 used_vcpus=20 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 923.453546] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.635528] env[62552]: DEBUG nova.compute.manager [req-47ed67a3-fa20-45c0-bd87-b7594f949e8d req-7196bbe2-2ac4-4421-a40c-a3763a88413c service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Received event network-vif-plugged-cd1b0e6a-5546-47af-84df-a5e05587638d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 923.635940] env[62552]: DEBUG oslo_concurrency.lockutils [req-47ed67a3-fa20-45c0-bd87-b7594f949e8d req-7196bbe2-2ac4-4421-a40c-a3763a88413c service nova] Acquiring lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.636226] env[62552]: DEBUG oslo_concurrency.lockutils [req-47ed67a3-fa20-45c0-bd87-b7594f949e8d req-7196bbe2-2ac4-4421-a40c-a3763a88413c service nova] Lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.636413] env[62552]: DEBUG oslo_concurrency.lockutils [req-47ed67a3-fa20-45c0-bd87-b7594f949e8d req-7196bbe2-2ac4-4421-a40c-a3763a88413c service nova] Lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.636591] env[62552]: DEBUG nova.compute.manager [req-47ed67a3-fa20-45c0-bd87-b7594f949e8d req-7196bbe2-2ac4-4421-a40c-a3763a88413c service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] No waiting events found dispatching network-vif-plugged-cd1b0e6a-5546-47af-84df-a5e05587638d {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 923.636768] env[62552]: WARNING nova.compute.manager [req-47ed67a3-fa20-45c0-bd87-b7594f949e8d req-7196bbe2-2ac4-4421-a40c-a3763a88413c service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Received unexpected event network-vif-plugged-cd1b0e6a-5546-47af-84df-a5e05587638d for instance with vm_state building and task_state spawning. [ 923.659869] env[62552]: DEBUG nova.compute.manager [req-6977ec0e-a986-4ae5-822e-35b268298ad9 req-ab0f05d5-b2b6-4a84-bd6f-2d045c1d9ac1 service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Received event network-changed-cd1b0e6a-5546-47af-84df-a5e05587638d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 923.660138] env[62552]: DEBUG nova.compute.manager [req-6977ec0e-a986-4ae5-822e-35b268298ad9 req-ab0f05d5-b2b6-4a84-bd6f-2d045c1d9ac1 service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Refreshing instance network info cache due to event network-changed-cd1b0e6a-5546-47af-84df-a5e05587638d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 923.660444] env[62552]: DEBUG oslo_concurrency.lockutils [req-6977ec0e-a986-4ae5-822e-35b268298ad9 req-ab0f05d5-b2b6-4a84-bd6f-2d045c1d9ac1 service nova] Acquiring lock "refresh_cache-ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.660609] env[62552]: DEBUG oslo_concurrency.lockutils [req-6977ec0e-a986-4ae5-822e-35b268298ad9 req-ab0f05d5-b2b6-4a84-bd6f-2d045c1d9ac1 service nova] Acquired lock "refresh_cache-ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.660809] env[62552]: DEBUG nova.network.neutron [req-6977ec0e-a986-4ae5-822e-35b268298ad9 req-ab0f05d5-b2b6-4a84-bd6f-2d045c1d9ac1 service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Refreshing network info cache for port cd1b0e6a-5546-47af-84df-a5e05587638d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.722476] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239706, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.753587] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.779193] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.855816} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.779193] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.779193] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73803f17-abba-4e28-a3f3-dc150bdabda2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.805198] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] e8e8dec1-4210-44d2-a3f8-76055f2c3b57/e8e8dec1-4210-44d2-a3f8-76055f2c3b57.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.807489] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7283be7-3a70-4491-8ea7-a70610bd7ebb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.826371] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c17561c-b6ae-49c5-b7eb-bd22038136a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.837127] env[62552]: DEBUG oslo_vmware.api [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239705, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.842861] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a2139a-33d3-4980-8c00-c60b94cb6608 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.851032] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.852410] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 923.852410] env[62552]: value = "task-1239708" [ 923.852410] env[62552]: _type = "Task" [ 923.852410] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.852748] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239707, 'name': PowerOffVM_Task, 'duration_secs': 0.471907} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.853639] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 923.857713] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0517d044-9220-457c-9e0e-936a15e5df00 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.887756] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2798b11-cea0-410b-aa0a-0e7cdf95a870 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.909910] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239708, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.913885] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb9584c-db7f-430b-a365-44e0017ccfd6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.921384] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04283db0-9f3e-4ef3-978d-d3f3b4301c39 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.926424] env[62552]: DEBUG nova.compute.manager [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 923.926561] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.927915] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a28ebc7-32f1-418b-88d0-9f31a924a918 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.951362] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "refresh_cache-ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.951876] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.956405] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.956994] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48cc2f7b-3630-4ff2-90a9-3b31be301f0b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.964843] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.972862] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.973529] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82bd1bbc-71c1-4e8e-88b2-6f818dd7f68e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.977326] env[62552]: DEBUG oslo_vmware.api [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 923.977326] env[62552]: value = "task-1239709" [ 923.977326] env[62552]: _type = "Task" [ 923.977326] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.985129] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 923.985129] env[62552]: value = "task-1239710" [ 923.985129] env[62552]: _type = "Task" [ 923.985129] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.992695] env[62552]: DEBUG oslo_vmware.api [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.000903] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 924.001171] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.001479] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.001634] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.001820] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.002106] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f15f2298-ee7b-4a7d-8565-05ed26f14ba4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.018357] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.018576] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 924.019544] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d60fb42-5d7b-439e-873d-28d3fb0040a5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.027410] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 924.027410] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522602d5-62ab-ab6f-c444-aee721c7c2ff" [ 924.027410] env[62552]: _type = "Task" [ 924.027410] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.037759] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522602d5-62ab-ab6f-c444-aee721c7c2ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.221629] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239706, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.233773] env[62552]: DEBUG nova.network.neutron [req-6977ec0e-a986-4ae5-822e-35b268298ad9 req-ab0f05d5-b2b6-4a84-bd6f-2d045c1d9ac1 service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 924.295948] env[62552]: DEBUG oslo_vmware.api [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239705, 'name': PowerOffVM_Task, 'duration_secs': 1.287999} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.296257] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.296438] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.296714] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8392dea2-cdd7-4deb-9d4e-b53d7ab6467d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.361832] env[62552]: DEBUG nova.compute.manager [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 924.362778] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384fe13a-bbe8-4d4e-b70e-bbefc00f6271 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.370097] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.370347] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.370544] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Deleting the datastore file [datastore2] 0080a6b3-cac0-44f3-a037-3f94f1daa275 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.373767] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-466ced03-2b2f-4612-8f89-abc1b1fde4ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.375721] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239708, 'name': ReconfigVM_Task, 'duration_secs': 0.370699} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.377493] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Reconfigured VM instance instance-00000051 to attach disk [datastore1] e8e8dec1-4210-44d2-a3f8-76055f2c3b57/e8e8dec1-4210-44d2-a3f8-76055f2c3b57.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.380944] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3a01d57-625e-426c-86ab-49c9947d5237 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.387079] env[62552]: DEBUG oslo_vmware.api [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for the task: (returnval){ [ 924.387079] env[62552]: value = "task-1239712" [ 924.387079] env[62552]: _type = "Task" [ 924.387079] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.391115] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 924.391115] env[62552]: value = "task-1239713" [ 924.391115] env[62552]: _type = "Task" [ 924.391115] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.398453] env[62552]: DEBUG oslo_vmware.api [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239712, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.403577] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239713, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.421377] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "34736dd0-e617-475e-baa2-cb372db1afb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.421706] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "34736dd0-e617-475e-baa2-cb372db1afb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.421942] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "34736dd0-e617-475e-baa2-cb372db1afb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.422218] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "34736dd0-e617-475e-baa2-cb372db1afb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.422370] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "34736dd0-e617-475e-baa2-cb372db1afb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.424729] env[62552]: INFO nova.compute.manager [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Terminating instance [ 924.440273] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.458571] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 924.491604] env[62552]: DEBUG oslo_vmware.api [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239709, 'name': PowerOffVM_Task, 'duration_secs': 0.278768} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.492179] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.492179] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.492448] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8cfdc182-9576-451c-9cd4-117be36285c9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.505501] env[62552]: DEBUG nova.network.neutron [req-6977ec0e-a986-4ae5-822e-35b268298ad9 req-ab0f05d5-b2b6-4a84-bd6f-2d045c1d9ac1 service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.540491] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522602d5-62ab-ab6f-c444-aee721c7c2ff, 'name': SearchDatastore_Task, 'duration_secs': 0.085465} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.541601] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5efdfdef-44c2-4663-bf44-365574ef9e84 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.549702] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 924.549702] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522f19d5-047d-d360-c531-92db27517b14" [ 924.549702] env[62552]: _type = "Task" [ 924.549702] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.562430] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522f19d5-047d-d360-c531-92db27517b14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.563881] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.564112] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.564326] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Deleting the datastore file [datastore2] 4083ee43-ecea-4ea5-8923-42b348893824 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.564603] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2d2a6bf-42c1-4218-9264-aae8c1e1b34f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.575304] env[62552]: DEBUG oslo_vmware.api [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 924.575304] env[62552]: value = "task-1239715" [ 924.575304] env[62552]: _type = "Task" [ 924.575304] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.587370] env[62552]: DEBUG oslo_vmware.api [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239715, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.730904] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239706, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.907360] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239713, 'name': Rename_Task, 'duration_secs': 0.161442} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.909337] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.909741] env[62552]: DEBUG oslo_vmware.api [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Task: {'id': task-1239712, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360903} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.909985] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12148ec9-3200-44c1-9d9e-ea0e85790a6e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.912053] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.912382] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 924.912505] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 924.912721] env[62552]: INFO nova.compute.manager [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Took 2.15 seconds to destroy the instance on the hypervisor. [ 924.912975] env[62552]: DEBUG oslo.service.loopingcall [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.913274] env[62552]: DEBUG nova.compute.manager [-] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 924.913375] env[62552]: DEBUG nova.network.neutron [-] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 924.925973] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 924.925973] env[62552]: value = "task-1239716" [ 924.925973] env[62552]: _type = "Task" [ 924.925973] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.932965] env[62552]: DEBUG nova.compute.manager [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 924.933281] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 924.940563] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd7942f-ef5f-4f98-ab0a-4bbfa6f5d77f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.953794] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.961566] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239716, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.962196] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.962967] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-356604b9-1fc9-47d0-9d34-c8dc90ccb1ad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.965360] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 924.965588] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.585s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.966241] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 20.682s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.979048] env[62552]: DEBUG oslo_vmware.api [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 924.979048] env[62552]: value = "task-1239717" [ 924.979048] env[62552]: _type = "Task" [ 924.979048] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.991251] env[62552]: DEBUG oslo_vmware.api [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239717, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.007898] env[62552]: DEBUG oslo_concurrency.lockutils [req-6977ec0e-a986-4ae5-822e-35b268298ad9 req-ab0f05d5-b2b6-4a84-bd6f-2d045c1d9ac1 service nova] Releasing lock "refresh_cache-ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.008474] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquired lock "refresh_cache-ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.008587] env[62552]: DEBUG nova.network.neutron [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.065237] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522f19d5-047d-d360-c531-92db27517b14, 'name': SearchDatastore_Task, 'duration_secs': 0.089452} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.065673] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.065986] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 11831580-1b58-476a-91ce-a4e55947fd91/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. {{(pid=62552) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 925.066304] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7fc2d03-943c-41d8-b688-3531675268c2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.076121] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 925.076121] env[62552]: value = "task-1239718" [ 925.076121] env[62552]: _type = "Task" [ 925.076121] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.092847] env[62552]: DEBUG oslo_vmware.api [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239715, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.39608} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.096496] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.096732] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.096912] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.097117] env[62552]: INFO nova.compute.manager [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Took 1.17 seconds to destroy the instance on the hypervisor. [ 925.097394] env[62552]: DEBUG oslo.service.loopingcall [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.097609] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.097861] env[62552]: DEBUG nova.compute.manager [-] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 925.098068] env[62552]: DEBUG nova.network.neutron [-] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.230617] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239706, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.395596] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882b076d-2ae1-4f93-9148-3d9febcc21ea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.410823] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Doing hard reboot of VM {{(pid=62552) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 925.414091] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-b1b727ea-9454-4dbf-aa15-239585f985bd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.421955] env[62552]: DEBUG oslo_vmware.api [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 925.421955] env[62552]: value = "task-1239719" [ 925.421955] env[62552]: _type = "Task" [ 925.421955] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.436785] env[62552]: DEBUG oslo_vmware.api [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239719, 'name': ResetVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.452512] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239716, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.462408] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.493927] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "c5ed17ee-7421-49d7-975f-1515de1e4b01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.494355] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "c5ed17ee-7421-49d7-975f-1515de1e4b01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.507329] env[62552]: DEBUG oslo_vmware.api [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239717, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.584527] env[62552]: DEBUG nova.network.neutron [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.601270] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.712782] env[62552]: DEBUG nova.compute.manager [req-fde1140a-6a48-4356-a111-0ad336bb19f8 req-e0e4e847-4cf7-44af-bba8-35149ed3faa0 service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Received event network-vif-deleted-3e0578b1-d107-4e54-8f04-1299abff83ef {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 925.713070] env[62552]: INFO nova.compute.manager [req-fde1140a-6a48-4356-a111-0ad336bb19f8 req-e0e4e847-4cf7-44af-bba8-35149ed3faa0 service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Neutron deleted interface 3e0578b1-d107-4e54-8f04-1299abff83ef; detaching it from the instance and deleting it from the info cache [ 925.713264] env[62552]: DEBUG nova.network.neutron [req-fde1140a-6a48-4356-a111-0ad336bb19f8 req-e0e4e847-4cf7-44af-bba8-35149ed3faa0 service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.729460] env[62552]: DEBUG nova.compute.manager [req-54494347-587b-49ab-a344-eb5047007f5e req-6992e173-7f6c-46d1-b1c0-bf9fe9a4f2c3 service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Received event network-vif-deleted-b77f3748-ccd8-4287-9ed2-55c69d3c783f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 925.729700] env[62552]: INFO nova.compute.manager [req-54494347-587b-49ab-a344-eb5047007f5e req-6992e173-7f6c-46d1-b1c0-bf9fe9a4f2c3 service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Neutron deleted interface b77f3748-ccd8-4287-9ed2-55c69d3c783f; detaching it from the instance and deleting it from the info cache [ 925.729903] env[62552]: DEBUG nova.network.neutron [req-54494347-587b-49ab-a344-eb5047007f5e req-6992e173-7f6c-46d1-b1c0-bf9fe9a4f2c3 service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.736341] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239706, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.856732] env[62552]: DEBUG nova.network.neutron [-] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.905590] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd463720-028b-412a-a193-1b793726df1f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.919128] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3543a418-64a1-4134-8bcd-21fe9533d299 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.964168] env[62552]: DEBUG nova.network.neutron [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Updating instance_info_cache with network_info: [{"id": "cd1b0e6a-5546-47af-84df-a5e05587638d", "address": "fa:16:3e:84:f9:c9", "network": {"id": "1657d27d-1308-40b2-be9a-a1dd2b1da7ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1466254829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adb0b1256dd34e33913d5ae5bb3553b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd1b0e6a-55", "ovs_interfaceid": "cd1b0e6a-5546-47af-84df-a5e05587638d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.968669] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24209797-2463-4497-b660-7ee79a45d0bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.972281] env[62552]: DEBUG oslo_vmware.api [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239716, 'name': PowerOnVM_Task, 'duration_secs': 0.74323} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.972281] env[62552]: DEBUG oslo_vmware.api [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239719, 'name': ResetVM_Task, 'duration_secs': 0.194214} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.972281] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.972484] env[62552]: INFO nova.compute.manager [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Took 10.13 seconds to spawn the instance on the hypervisor. [ 925.972663] env[62552]: DEBUG nova.compute.manager [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 925.972951] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Did hard reboot of VM {{(pid=62552) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 925.973151] env[62552]: DEBUG nova.compute.manager [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 925.974781] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553701ec-a77f-44a5-8fa4-83229da68137 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.978491] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d1d60e-9b4e-4741-9c63-310662c5f682 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.987042] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.991833] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c1ef3a-ab4c-4616-b39c-532579ab3382 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.999176] env[62552]: DEBUG nova.compute.manager [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 926.024405] env[62552]: DEBUG nova.compute.provider_tree [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.024405] env[62552]: DEBUG oslo_vmware.api [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239717, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.092856] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.201859] env[62552]: DEBUG nova.network.neutron [-] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.219972] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9232acd6-1ce0-4319-806c-d2cfb7f70312 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.226859] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239706, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.811593} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.227910] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d88505cd-62b9-4da3-926e-217d15bab510/d88505cd-62b9-4da3-926e-217d15bab510.vmdk to [datastore1] cdee22a9-4327-47af-ab14-8403a06b2802/cdee22a9-4327-47af-ab14-8403a06b2802.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 926.228718] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3dee37-78f4-43b0-b92e-ebd7c7136c0e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.235374] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df01d1a-7f3c-4b97-ba35-d08874225144 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.246332] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4574be2b-8fc6-4792-aae2-6ac509626e95 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.266317] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] cdee22a9-4327-47af-ab14-8403a06b2802/cdee22a9-4327-47af-ab14-8403a06b2802.vmdk or device None with type streamOptimized {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.278776] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4461063f-fb41-45df-8ceb-303a5d8b66fc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.293270] env[62552]: DEBUG nova.compute.manager [req-fde1140a-6a48-4356-a111-0ad336bb19f8 req-e0e4e847-4cf7-44af-bba8-35149ed3faa0 service nova] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Detach interface failed, port_id=3e0578b1-d107-4e54-8f04-1299abff83ef, reason: Instance 0080a6b3-cac0-44f3-a037-3f94f1daa275 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 926.298329] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c43194b-8f56-48fb-92e9-e450cab9a817 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.309183] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 926.309183] env[62552]: value = "task-1239720" [ 926.309183] env[62552]: _type = "Task" [ 926.309183] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.317341] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239720, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.329288] env[62552]: DEBUG nova.compute.manager [req-54494347-587b-49ab-a344-eb5047007f5e req-6992e173-7f6c-46d1-b1c0-bf9fe9a4f2c3 service nova] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Detach interface failed, port_id=b77f3748-ccd8-4287-9ed2-55c69d3c783f, reason: Instance 4083ee43-ecea-4ea5-8923-42b348893824 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 926.359129] env[62552]: INFO nova.compute.manager [-] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Took 1.45 seconds to deallocate network for instance. [ 926.478814] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Releasing lock "refresh_cache-ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.478814] env[62552]: DEBUG nova.compute.manager [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Instance network_info: |[{"id": "cd1b0e6a-5546-47af-84df-a5e05587638d", "address": "fa:16:3e:84:f9:c9", "network": {"id": "1657d27d-1308-40b2-be9a-a1dd2b1da7ab", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1466254829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adb0b1256dd34e33913d5ae5bb3553b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd1b0e6a-55", "ovs_interfaceid": "cd1b0e6a-5546-47af-84df-a5e05587638d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 926.478814] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.479501] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:f9:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06cc7c49-c46c-4c1e-bf51-77e9ea802c40', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cd1b0e6a-5546-47af-84df-a5e05587638d', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.487404] env[62552]: DEBUG oslo.service.loopingcall [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.487733] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.491105] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1f20a91-112e-4f2f-b499-c0b8375bb508 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.520026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a241570e-ca2f-478c-a47e-f9982eb7547d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.452s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.529963] env[62552]: DEBUG nova.scheduler.client.report [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 926.539967] env[62552]: DEBUG oslo_vmware.api [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239717, 'name': PowerOffVM_Task, 'duration_secs': 1.231444} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.541662] env[62552]: INFO nova.compute.manager [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Took 46.29 seconds to build instance. [ 926.544274] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.544274] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.544654] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.544654] env[62552]: value = "task-1239721" [ 926.544654] env[62552]: _type = "Task" [ 926.544654] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.544966] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3cc365b-5cf8-469e-a422-1b967e9eb508 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.553289] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.560098] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239721, 'name': CreateVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.603214] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239718, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.638037] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 926.638037] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 926.638037] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleting the datastore file [datastore1] 34736dd0-e617-475e-baa2-cb372db1afb2 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.638037] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b1d96e9-ebeb-4a5a-8057-aea0f45ad446 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.646180] env[62552]: DEBUG oslo_vmware.api [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 926.646180] env[62552]: value = "task-1239723" [ 926.646180] env[62552]: _type = "Task" [ 926.646180] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.657278] env[62552]: DEBUG oslo_vmware.api [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.704782] env[62552]: INFO nova.compute.manager [-] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Took 1.61 seconds to deallocate network for instance. [ 926.822576] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239720, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.865661] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.968519] env[62552]: DEBUG oslo_vmware.api [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239699, 'name': ReconfigVM_Task, 'duration_secs': 5.841264} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.968789] env[62552]: DEBUG oslo_concurrency.lockutils [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.969015] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Reconfigured VM to detach interface {{(pid=62552) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 927.048997] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c18a13f0-d71b-47ad-b1ca-81e76751eb96 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.224s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.058533] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239721, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.096791] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239718, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.778953} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.097171] env[62552]: INFO nova.virt.vmwareapi.ds_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 11831580-1b58-476a-91ce-a4e55947fd91/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. [ 927.098051] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36507067-6bfa-4889-8744-b8b46cdfb32c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.123968] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 11831580-1b58-476a-91ce-a4e55947fd91/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.124614] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31a4ae9a-71a4-4249-800c-df6f7ab102d3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.143680] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 927.143680] env[62552]: value = "task-1239724" [ 927.143680] env[62552]: _type = "Task" [ 927.143680] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.154296] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239724, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.157260] env[62552]: DEBUG oslo_vmware.api [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.213207] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.328838] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239720, 'name': ReconfigVM_Task, 'duration_secs': 0.596381} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.329164] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Reconfigured VM instance instance-0000002d to attach disk [datastore1] cdee22a9-4327-47af-ab14-8403a06b2802/cdee22a9-4327-47af-ab14-8403a06b2802.vmdk or device None with type streamOptimized {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.331651] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'guest_format': None, 'encryption_options': None, 'device_type': 'disk', 'size': 0, 'disk_bus': None, 'boot_index': 0, 'encryption_secret_uuid': None, 'encrypted': False, 'encryption_format': None, 'image_id': '562b48de-e081-4c62-a3e2-cf4ccfd0d1ee'}], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'mount_device': '/dev/sdb', 'delete_on_termination': False, 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267441', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'name': 'volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cdee22a9-4327-47af-ab14-8403a06b2802', 'attached_at': '', 'detached_at': '', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'serial': '250ad8c9-818e-4ebf-864a-a2e15cd300d3'}, 'disk_bus': None, 'boot_index': None, 'attachment_id': '1de62591-fd4b-4df0-86f6-b62bde86bbc9', 'volume_type': None}], 'swap': None} {{(pid=62552) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 927.331880] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 927.332098] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267441', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'name': 'volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cdee22a9-4327-47af-ab14-8403a06b2802', 'attached_at': '', 'detached_at': '', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'serial': '250ad8c9-818e-4ebf-864a-a2e15cd300d3'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 927.333295] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8922f638-bffb-470c-83a4-957cc9703775 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.350447] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501bc6b6-65cf-4423-86c0-b01d784ef6d4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.379518] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3/volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.379857] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9fc6a82-e878-4f09-941f-c941c45a4b10 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.400035] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 927.400035] env[62552]: value = "task-1239725" [ 927.400035] env[62552]: _type = "Task" [ 927.400035] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.407930] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239725, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.550738] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.584s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.553828] env[62552]: DEBUG oslo_concurrency.lockutils [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.865s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.554060] env[62552]: DEBUG oslo_concurrency.lockutils [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.556071] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.495s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.556276] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.558597] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.805s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.558871] env[62552]: DEBUG nova.objects.instance [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lazy-loading 'resources' on Instance uuid fb881c34-dccc-4703-af8e-c75caafd9b08 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.571088] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239721, 'name': CreateVM_Task, 'duration_secs': 0.883433} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.572204] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.573032] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.573295] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.573683] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.574314] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39931f91-0a05-4c55-836b-2b4cfe8112fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.579771] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 927.579771] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52afac32-c58a-1620-9052-f706e0d08504" [ 927.579771] env[62552]: _type = "Task" [ 927.579771] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.589206] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52afac32-c58a-1620-9052-f706e0d08504, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.596460] env[62552]: INFO nova.scheduler.client.report [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Deleted allocations for instance 5cee4242-49ae-4cb7-a208-e2982f52fbad [ 927.608702] env[62552]: INFO nova.scheduler.client.report [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Deleted allocations for instance 43d1e732-7e9b-4f9f-b67c-f0ef2be91902 [ 927.657958] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239724, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.664443] env[62552]: DEBUG oslo_vmware.api [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239723, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.631074} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.664751] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.664963] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.665184] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.665390] env[62552]: INFO nova.compute.manager [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Took 2.73 seconds to destroy the instance on the hypervisor. [ 927.666123] env[62552]: DEBUG oslo.service.loopingcall [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.666123] env[62552]: DEBUG nova.compute.manager [-] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 927.666123] env[62552]: DEBUG nova.network.neutron [-] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.876497] env[62552]: DEBUG nova.compute.manager [req-a3835f7e-e473-4dc8-ae0c-ce7d01391fb0 req-7b3ad1e4-e1e3-44ad-ac2d-5c77820149e4 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-vif-deleted-035ebcb4-aada-41ce-9361-47d4162965cf {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 927.876816] env[62552]: INFO nova.compute.manager [req-a3835f7e-e473-4dc8-ae0c-ce7d01391fb0 req-7b3ad1e4-e1e3-44ad-ac2d-5c77820149e4 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Neutron deleted interface 035ebcb4-aada-41ce-9361-47d4162965cf; detaching it from the instance and deleting it from the info cache [ 927.877163] env[62552]: DEBUG nova.network.neutron [req-a3835f7e-e473-4dc8-ae0c-ce7d01391fb0 req-7b3ad1e4-e1e3-44ad-ac2d-5c77820149e4 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [{"id": "0679d99d-6c79-408f-82a1-9f80978a1c48", "address": "fa:16:3e:c1:12:aa", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0679d99d-6c", "ovs_interfaceid": "0679d99d-6c79-408f-82a1-9f80978a1c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9726e936-f144-4cce-9964-78bcae1c40ae", "address": "fa:16:3e:28:e5:36", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9726e936-f1", "ovs_interfaceid": "9726e936-f144-4cce-9964-78bcae1c40ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.914706] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239725, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.095766] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52afac32-c58a-1620-9052-f706e0d08504, 'name': SearchDatastore_Task, 'duration_secs': 0.010452} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.101613] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.101822] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.102332] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.102494] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.102693] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.103630] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb475fb4-6007-4331-9665-0772d18a8edd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.109192] env[62552]: DEBUG oslo_concurrency.lockutils [None req-765ffce1-b76b-4819-a677-cc7b6919616a tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "5cee4242-49ae-4cb7-a208-e2982f52fbad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.178s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.122676] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.122966] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 928.124203] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c3aa145f-98d6-45a4-aeac-ce2232683836 tempest-ServerPasswordTestJSON-407297882 tempest-ServerPasswordTestJSON-407297882-project-member] Lock "43d1e732-7e9b-4f9f-b67c-f0ef2be91902" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.627s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.126218] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d97f923-879b-42ac-9a39-b8fe6d39f874 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.133477] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 928.133477] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524ba543-3407-a26b-fa72-2d2b79a2a789" [ 928.133477] env[62552]: _type = "Task" [ 928.133477] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.148817] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524ba543-3407-a26b-fa72-2d2b79a2a789, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.159976] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239724, 'name': ReconfigVM_Task, 'duration_secs': 0.525836} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.161028] env[62552]: INFO nova.scheduler.client.report [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted allocation for migration ac053045-ef61-4508-bfae-a9c3df4952ed [ 928.162186] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 11831580-1b58-476a-91ce-a4e55947fd91/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.163648] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b9f414-e420-443c-a407-0a7a41589cc3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.199033] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f4c5d90-e127-4392-947d-25a7b386b024 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.220679] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 928.220679] env[62552]: value = "task-1239726" [ 928.220679] env[62552]: _type = "Task" [ 928.220679] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.232512] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239726, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.379823] env[62552]: DEBUG oslo_concurrency.lockutils [req-a3835f7e-e473-4dc8-ae0c-ce7d01391fb0 req-7b3ad1e4-e1e3-44ad-ac2d-5c77820149e4 service nova] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.379823] env[62552]: DEBUG oslo_concurrency.lockutils [req-a3835f7e-e473-4dc8-ae0c-ce7d01391fb0 req-7b3ad1e4-e1e3-44ad-ac2d-5c77820149e4 service nova] Acquired lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.380973] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a99c24b-4d89-4e9a-99cd-f5ddf519e7e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.399097] env[62552]: DEBUG oslo_concurrency.lockutils [req-a3835f7e-e473-4dc8-ae0c-ce7d01391fb0 req-7b3ad1e4-e1e3-44ad-ac2d-5c77820149e4 service nova] Releasing lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.399435] env[62552]: WARNING nova.compute.manager [req-a3835f7e-e473-4dc8-ae0c-ce7d01391fb0 req-7b3ad1e4-e1e3-44ad-ac2d-5c77820149e4 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Detach interface failed, port_id=035ebcb4-aada-41ce-9361-47d4162965cf, reason: No device with interface-id 035ebcb4-aada-41ce-9361-47d4162965cf exists on VM: nova.exception.NotFound: No device with interface-id 035ebcb4-aada-41ce-9361-47d4162965cf exists on VM [ 928.413851] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239725, 'name': ReconfigVM_Task, 'duration_secs': 0.974069} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.415831] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Reconfigured VM instance instance-0000002d to attach disk [datastore1] volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3/volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.424120] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f253ed91-dee6-4aab-97f0-1d49a757731d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.438743] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 928.438743] env[62552]: value = "task-1239727" [ 928.438743] env[62552]: _type = "Task" [ 928.438743] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.447261] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239727, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.462146] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e157b0ea-5fa0-48e0-b9c0-612ebba8f719 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.471425] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfda4d7-da8f-42a3-921f-9982152f7529 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.503403] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eae71fc-be74-4e50-b59e-5c45505226d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.510335] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.510639] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.510838] env[62552]: INFO nova.compute.manager [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Shelving [ 928.514722] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f7d2a4-67e5-4867-b2d9-ebf168ff4828 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.528202] env[62552]: DEBUG nova.compute.provider_tree [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.556570] env[62552]: DEBUG nova.network.neutron [-] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.643115] env[62552]: DEBUG oslo_concurrency.lockutils [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.643349] env[62552]: DEBUG oslo_concurrency.lockutils [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.643588] env[62552]: DEBUG nova.network.neutron [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 928.649047] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524ba543-3407-a26b-fa72-2d2b79a2a789, 'name': SearchDatastore_Task, 'duration_secs': 0.013647} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.649964] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-391a47f5-2348-4779-9b3b-adc6e19b0d25 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.655911] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 928.655911] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a1994f-9446-8d90-310d-c0cc6c6dc890" [ 928.655911] env[62552]: _type = "Task" [ 928.655911] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.664621] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a1994f-9446-8d90-310d-c0cc6c6dc890, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.670661] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a5a05db6-6cb7-4f61-8ed0-c30a60e53e4c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 28.106s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.717152] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "714fb65d-9f80-4a81-a637-3e4398405d9b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.717152] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "714fb65d-9f80-4a81-a637-3e4398405d9b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.717152] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "714fb65d-9f80-4a81-a637-3e4398405d9b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.717289] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "714fb65d-9f80-4a81-a637-3e4398405d9b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.717420] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "714fb65d-9f80-4a81-a637-3e4398405d9b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.719516] env[62552]: INFO nova.compute.manager [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Terminating instance [ 928.730517] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239726, 'name': ReconfigVM_Task, 'duration_secs': 0.236894} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.730792] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.731044] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d14abe59-0b68-4f32-872e-b4bb7eec85bf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.738200] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 928.738200] env[62552]: value = "task-1239728" [ 928.738200] env[62552]: _type = "Task" [ 928.738200] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.746511] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239728, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.950075] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.031526] env[62552]: DEBUG nova.scheduler.client.report [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 929.058117] env[62552]: INFO nova.compute.manager [-] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Took 1.39 seconds to deallocate network for instance. [ 929.167448] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a1994f-9446-8d90-310d-c0cc6c6dc890, 'name': SearchDatastore_Task, 'duration_secs': 0.025897} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.167769] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.168049] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] ae517e5f-0cd1-457c-99a6-6cd46fdd6b06/ae517e5f-0cd1-457c-99a6-6cd46fdd6b06.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.168378] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4a2f084-c32c-49d2-bf61-d75337b87ea1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.175734] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 929.175734] env[62552]: value = "task-1239729" [ 929.175734] env[62552]: _type = "Task" [ 929.175734] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.185250] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239729, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.228397] env[62552]: DEBUG nova.compute.manager [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 929.228918] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.230032] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5542931c-d0a7-4b8a-9c37-8ce5c5cdc85c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.239484] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.242993] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a0b187c-2ae5-45db-bb13-c8212668fb6d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.257830] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239728, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.258227] env[62552]: DEBUG oslo_vmware.api [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 929.258227] env[62552]: value = "task-1239730" [ 929.258227] env[62552]: _type = "Task" [ 929.258227] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.269481] env[62552]: DEBUG oslo_vmware.api [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.451244] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239727, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.472736] env[62552]: DEBUG oslo_concurrency.lockutils [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.472985] env[62552]: DEBUG oslo_concurrency.lockutils [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.473229] env[62552]: DEBUG oslo_concurrency.lockutils [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.473437] env[62552]: DEBUG oslo_concurrency.lockutils [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.473632] env[62552]: DEBUG oslo_concurrency.lockutils [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.475952] env[62552]: INFO nova.compute.manager [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Terminating instance [ 929.525609] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.526190] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5bfd96b4-32c4-4815-9866-a1f7bd2bda29 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.537949] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.979s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.550994] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.996s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.550994] env[62552]: INFO nova.compute.claims [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.553944] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 929.553944] env[62552]: value = "task-1239731" [ 929.553944] env[62552]: _type = "Task" [ 929.553944] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.566212] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.566594] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.573355] env[62552]: INFO nova.scheduler.client.report [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Deleted allocations for instance fb881c34-dccc-4703-af8e-c75caafd9b08 [ 929.658278] env[62552]: INFO nova.network.neutron [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Port 9726e936-f144-4cce-9964-78bcae1c40ae from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 929.658700] env[62552]: DEBUG nova.network.neutron [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [{"id": "0679d99d-6c79-408f-82a1-9f80978a1c48", "address": "fa:16:3e:c1:12:aa", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0679d99d-6c", "ovs_interfaceid": "0679d99d-6c79-408f-82a1-9f80978a1c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.688959] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239729, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.697128] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.697709] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.752489] env[62552]: DEBUG oslo_vmware.api [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239728, 'name': PowerOnVM_Task, 'duration_secs': 0.659439} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.752796] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.755765] env[62552]: DEBUG nova.compute.manager [None req-a7258142-a5a7-492d-b071-21409d5936d7 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 929.756700] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6faf0492-bbc9-4b0f-bfce-6fc260906575 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.772516] env[62552]: DEBUG oslo_vmware.api [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239730, 'name': PowerOffVM_Task, 'duration_secs': 0.459575} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.774380] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.774571] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 929.781462] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da5a8903-19a6-424f-bd6d-112298a9ba70 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.888281] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.888281] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.888281] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Deleting the datastore file [datastore1] 714fb65d-9f80-4a81-a637-3e4398405d9b {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.888281] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a56eff6d-2ec7-409b-b800-0a0c4f82397c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.896244] env[62552]: DEBUG oslo_vmware.api [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for the task: (returnval){ [ 929.896244] env[62552]: value = "task-1239733" [ 929.896244] env[62552]: _type = "Task" [ 929.896244] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.907923] env[62552]: DEBUG oslo_vmware.api [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239733, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.952643] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239727, 'name': ReconfigVM_Task, 'duration_secs': 1.012823} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.952700] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267441', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'name': 'volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cdee22a9-4327-47af-ab14-8403a06b2802', 'attached_at': '', 'detached_at': '', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'serial': '250ad8c9-818e-4ebf-864a-a2e15cd300d3'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 929.953392] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccd1e147-bb2f-4821-9083-b471d2860cf5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.960140] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 929.960140] env[62552]: value = "task-1239734" [ 929.960140] env[62552]: _type = "Task" [ 929.960140] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.970670] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239734, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.979638] env[62552]: DEBUG nova.compute.manager [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 929.979925] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.980882] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30eee063-265b-4dea-868b-20a868e675aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.987854] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.988144] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3752a8f-c4c6-4305-a3f9-a4a4b29502e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.996039] env[62552]: DEBUG oslo_vmware.api [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 929.996039] env[62552]: value = "task-1239735" [ 929.996039] env[62552]: _type = "Task" [ 929.996039] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.008359] env[62552]: DEBUG oslo_vmware.api [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.068594] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239731, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.082680] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82f4a378-edb1-4e80-99e7-7cd69bcf416a tempest-ServerShowV247Test-1613308735 tempest-ServerShowV247Test-1613308735-project-member] Lock "fb881c34-dccc-4703-af8e-c75caafd9b08" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.570s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.162135] env[62552]: DEBUG nova.compute.manager [req-87aded32-47d0-42c9-a43f-649cf8c55c09 req-c879b684-e9c3-4376-bbc9-c555c2c01ec1 service nova] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Received event network-vif-deleted-1e51c71e-8255-4974-a13e-53631ef167af {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 930.162426] env[62552]: DEBUG nova.compute.manager [req-87aded32-47d0-42c9-a43f-649cf8c55c09 req-c879b684-e9c3-4376-bbc9-c555c2c01ec1 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-vif-deleted-9726e936-f144-4cce-9964-78bcae1c40ae {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 930.165474] env[62552]: DEBUG oslo_concurrency.lockutils [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-e31e3077-04e1-4adb-a0cc-44cf84a89eda" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.187583] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239729, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.639453} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.187583] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] ae517e5f-0cd1-457c-99a6-6cd46fdd6b06/ae517e5f-0cd1-457c-99a6-6cd46fdd6b06.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 930.187794] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.188021] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-074f520d-6570-4239-9f94-223713ea8272 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.194411] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 930.194411] env[62552]: value = "task-1239736" [ 930.194411] env[62552]: _type = "Task" [ 930.194411] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.203619] env[62552]: DEBUG nova.compute.utils [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.209240] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239736, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.406909] env[62552]: DEBUG oslo_vmware.api [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239733, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.470149] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239734, 'name': Rename_Task, 'duration_secs': 0.177378} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.470475] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.470722] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc8cc129-6bfb-49c8-8593-fa4b9e89b464 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.476265] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 930.476265] env[62552]: value = "task-1239737" [ 930.476265] env[62552]: _type = "Task" [ 930.476265] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.483829] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239737, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.504093] env[62552]: DEBUG oslo_vmware.api [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239735, 'name': PowerOffVM_Task, 'duration_secs': 0.207172} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.504414] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.504584] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.504853] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6790836c-0f92-4c7a-a896-6eec057ec1c3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.571459] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239731, 'name': PowerOffVM_Task, 'duration_secs': 0.559769} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.571903] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.572599] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ce661e-cab6-4f6b-a468-43754207023c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.596599] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e64f202-e2b8-45b6-99bb-5ed9ebfea749 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.669643] env[62552]: DEBUG oslo_concurrency.lockutils [None req-472c8d10-960e-496c-873b-91567a4a15c1 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-e31e3077-04e1-4adb-a0cc-44cf84a89eda-035ebcb4-aada-41ce-9361-47d4162965cf" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.347s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.706532] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239736, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.22109} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.706798] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.707612] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6284ed64-901c-418b-9d80-808456e21bd4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.712770] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.015s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.734556] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] ae517e5f-0cd1-457c-99a6-6cd46fdd6b06/ae517e5f-0cd1-457c-99a6-6cd46fdd6b06.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.738141] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab29f8c0-8aa2-44dc-b1db-0bd4368253af {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.755446] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.755446] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.755446] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleting the datastore file [datastore2] e31e3077-04e1-4adb-a0cc-44cf84a89eda {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.755790] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1dae7698-aff8-477f-a013-b01c8148a289 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.762205] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 930.762205] env[62552]: value = "task-1239739" [ 930.762205] env[62552]: _type = "Task" [ 930.762205] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.765558] env[62552]: DEBUG oslo_vmware.api [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 930.765558] env[62552]: value = "task-1239740" [ 930.765558] env[62552]: _type = "Task" [ 930.765558] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.777735] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239739, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.781484] env[62552]: DEBUG oslo_vmware.api [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.903990] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebaa508d-5515-4479-a3a4-707ca91aeed2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.915269] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771e9f8f-9d5f-4e64-96cd-be84c092c54c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.918731] env[62552]: DEBUG oslo_vmware.api [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Task: {'id': task-1239733, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.535245} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.919103] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.919363] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 930.919579] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 930.920070] env[62552]: INFO nova.compute.manager [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Took 1.69 seconds to destroy the instance on the hypervisor. [ 930.920070] env[62552]: DEBUG oslo.service.loopingcall [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.920631] env[62552]: DEBUG nova.compute.manager [-] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 930.920747] env[62552]: DEBUG nova.network.neutron [-] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 930.953597] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b7fd8b-e65d-4c08-a015-f13ee02b598e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.961164] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3f91ce-04c5-4fff-9b48-931d6cab850c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.975183] env[62552]: DEBUG nova.compute.provider_tree [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.985852] env[62552]: DEBUG oslo_vmware.api [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239737, 'name': PowerOnVM_Task, 'duration_secs': 0.486129} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.986332] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 931.011477] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "22357d4e-9771-477c-9fc3-fe3d76f6e902" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.011646] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.011835] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "22357d4e-9771-477c-9fc3-fe3d76f6e902-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.012192] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.012192] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.014499] env[62552]: INFO nova.compute.manager [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Terminating instance [ 931.104197] env[62552]: DEBUG nova.compute.manager [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 931.104197] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2829c1-040c-4fe7-b1e9-f59719b317b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.106657] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 931.107060] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2b69c520-377a-4d63-bc4e-37c4636b43bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.117095] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 931.117095] env[62552]: value = "task-1239741" [ 931.117095] env[62552]: _type = "Task" [ 931.117095] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.127824] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239741, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.286051] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239739, 'name': ReconfigVM_Task, 'duration_secs': 0.423293} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.286523] env[62552]: DEBUG oslo_vmware.api [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259112} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.286929] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Reconfigured VM instance instance-00000052 to attach disk [datastore1] ae517e5f-0cd1-457c-99a6-6cd46fdd6b06/ae517e5f-0cd1-457c-99a6-6cd46fdd6b06.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.287752] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.288088] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.288419] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.288784] env[62552]: INFO nova.compute.manager [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Took 1.31 seconds to destroy the instance on the hypervisor. [ 931.289222] env[62552]: DEBUG oslo.service.loopingcall [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.289635] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b56cfa5-c0af-4c9f-9923-88b32fcf3f28 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.293287] env[62552]: DEBUG nova.compute.manager [-] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 931.293632] env[62552]: DEBUG nova.network.neutron [-] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.311307] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 931.311307] env[62552]: value = "task-1239742" [ 931.311307] env[62552]: _type = "Task" [ 931.311307] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.327824] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239742, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.392798] env[62552]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 9726e936-f144-4cce-9964-78bcae1c40ae could not be found.", "detail": ""}} {{(pid=62552) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 931.393141] env[62552]: DEBUG nova.network.neutron [-] Unable to show port 9726e936-f144-4cce-9964-78bcae1c40ae as it no longer exists. {{(pid=62552) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 931.478745] env[62552]: DEBUG nova.scheduler.client.report [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 931.518545] env[62552]: DEBUG nova.compute.manager [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 931.518804] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.519716] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c941f3-7861-4ebd-bc5e-55ea357a6f76 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.528226] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.528520] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30197902-00a4-492d-9916-d00491e0f75a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.535686] env[62552]: DEBUG oslo_vmware.api [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 931.535686] env[62552]: value = "task-1239743" [ 931.535686] env[62552]: _type = "Task" [ 931.535686] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.551209] env[62552]: DEBUG oslo_vmware.api [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.634444] env[62552]: DEBUG oslo_concurrency.lockutils [None req-029731fa-0e38-4193-9ebb-9e5dc5e213de tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 63.368s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.640303] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239741, 'name': CreateSnapshot_Task, 'duration_secs': 0.486951} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.640548] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 931.642441] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a2c15d-62a1-419b-b055-c984e623bf4d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.805631] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.807090] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.807090] env[62552]: INFO nova.compute.manager [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Attaching volume 76baa0b3-5dd2-4828-8f24-0aa4063096c8 to /dev/sdb [ 931.813018] env[62552]: DEBUG nova.network.neutron [-] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.827989] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239742, 'name': Rename_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.873338] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9803f3b0-fd26-4f52-b0ec-9fbbe7120f31 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.880948] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1916d876-bcbe-4974-a4b6-c47cc49d002d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.888207] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "04186fde-a9ed-415a-bd40-312f0347fcc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.888438] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.905029] env[62552]: DEBUG nova.virt.block_device [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updating existing volume attachment record: 80fd87ae-a233-4354-889c-048e1f420c72 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 931.985733] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.986514] env[62552]: DEBUG nova.compute.manager [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 931.990872] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.125s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.990872] env[62552]: DEBUG nova.objects.instance [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lazy-loading 'resources' on Instance uuid 0080a6b3-cac0-44f3-a037-3f94f1daa275 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.046894] env[62552]: DEBUG oslo_vmware.api [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.120565] env[62552]: DEBUG nova.compute.manager [req-b29dab4a-fc68-47c0-bdc1-f0377701b0c5 req-0dbcef1b-2600-4945-b062-94e3c8a63a26 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Received event network-vif-deleted-0679d99d-6c79-408f-82a1-9f80978a1c48 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 932.120838] env[62552]: INFO nova.compute.manager [req-b29dab4a-fc68-47c0-bdc1-f0377701b0c5 req-0dbcef1b-2600-4945-b062-94e3c8a63a26 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Neutron deleted interface 0679d99d-6c79-408f-82a1-9f80978a1c48; detaching it from the instance and deleting it from the info cache [ 932.121107] env[62552]: DEBUG nova.network.neutron [req-b29dab4a-fc68-47c0-bdc1-f0377701b0c5 req-0dbcef1b-2600-4945-b062-94e3c8a63a26 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.149890] env[62552]: INFO nova.compute.manager [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Unrescuing [ 932.149890] env[62552]: DEBUG oslo_concurrency.lockutils [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.149890] env[62552]: DEBUG oslo_concurrency.lockutils [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.149890] env[62552]: DEBUG nova.network.neutron [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.163562] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 932.164775] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2ae206c2-98e9-48bf-bae1-9d34adaa66cb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.173353] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 932.173353] env[62552]: value = "task-1239747" [ 932.173353] env[62552]: _type = "Task" [ 932.173353] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.181521] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239747, 'name': CloneVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.185103] env[62552]: DEBUG nova.compute.manager [req-384cd8ca-e230-4905-a671-6825133046b1 req-a3903f5b-2a17-44c0-8652-d258a5060aac service nova] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Received event network-vif-deleted-71e835ff-cc35-4e84-923d-1931fdb1c527 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 932.319561] env[62552]: INFO nova.compute.manager [-] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Took 1.40 seconds to deallocate network for instance. [ 932.329103] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239742, 'name': Rename_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.391065] env[62552]: DEBUG nova.compute.manager [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 932.419856] env[62552]: DEBUG nova.network.neutron [-] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.494954] env[62552]: DEBUG nova.compute.utils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 932.497719] env[62552]: DEBUG nova.compute.manager [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 932.498036] env[62552]: DEBUG nova.network.neutron [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 932.552074] env[62552]: DEBUG oslo_vmware.api [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.553461] env[62552]: DEBUG nova.policy [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '384d6877b7384d21b12a155d13056a7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f13a11803d1b44178f053cc0a4affedd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.626792] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6cf45755-b1f5-454a-95e6-03af45a478fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.638786] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c1cecb-7438-4f3a-a151-d1811b7395ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.672517] env[62552]: DEBUG nova.compute.manager [req-b29dab4a-fc68-47c0-bdc1-f0377701b0c5 req-0dbcef1b-2600-4945-b062-94e3c8a63a26 service nova] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Detach interface failed, port_id=0679d99d-6c79-408f-82a1-9f80978a1c48, reason: Instance e31e3077-04e1-4adb-a0cc-44cf84a89eda could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 932.682342] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239747, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.827482] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.833546] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239742, 'name': Rename_Task, 'duration_secs': 1.164127} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.834402] env[62552]: DEBUG nova.network.neutron [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Successfully created port: 8018d47b-da5e-40c6-b217-a8b685b44983 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.836892] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 932.846277] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-809e353f-05ce-4896-b961-4e489c0ea3a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.853761] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 932.853761] env[62552]: value = "task-1239748" [ 932.853761] env[62552]: _type = "Task" [ 932.853761] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.865772] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239748, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.868131] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49282164-937a-40cd-95f3-4ef095ba4898 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.876840] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5630af-75ba-47a6-86fc-0d01d3886d98 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.918334] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a6a3ef-1940-4ba7-9c49-73a07427557c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.926022] env[62552]: INFO nova.compute.manager [-] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Took 1.63 seconds to deallocate network for instance. [ 932.933516] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6a428b-b770-4745-a98e-ffefaedebacd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.939278] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.950390] env[62552]: DEBUG nova.compute.provider_tree [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.999097] env[62552]: DEBUG nova.network.neutron [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Updating instance_info_cache with network_info: [{"id": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "address": "fa:16:3e:ce:cc:01", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap123341c8-40", "ovs_interfaceid": "123341c8-40fb-4f81-9d1c-1f556b6a92b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.005134] env[62552]: DEBUG nova.compute.manager [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 933.047713] env[62552]: DEBUG oslo_vmware.api [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239743, 'name': PowerOffVM_Task, 'duration_secs': 1.164535} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.047992] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.048186] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.048436] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79a57a16-c8f1-41ad-83e1-7ca0387aa885 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.126630] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.126895] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.127161] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleting the datastore file [datastore2] 22357d4e-9771-477c-9fc3-fe3d76f6e902 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.127446] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c2087f11-4ed2-483a-b4ef-dcbef3249530 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.133089] env[62552]: DEBUG oslo_vmware.api [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 933.133089] env[62552]: value = "task-1239750" [ 933.133089] env[62552]: _type = "Task" [ 933.133089] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.141518] env[62552]: DEBUG oslo_vmware.api [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239750, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.185745] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239747, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.363527] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239748, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.440765] env[62552]: DEBUG oslo_concurrency.lockutils [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.457263] env[62552]: DEBUG nova.scheduler.client.report [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 933.501923] env[62552]: DEBUG oslo_concurrency.lockutils [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "refresh_cache-11831580-1b58-476a-91ce-a4e55947fd91" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.502836] env[62552]: DEBUG nova.objects.instance [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lazy-loading 'flavor' on Instance uuid 11831580-1b58-476a-91ce-a4e55947fd91 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.643506] env[62552]: DEBUG oslo_vmware.api [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239750, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.3666} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.643868] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.644101] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.644292] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.644491] env[62552]: INFO nova.compute.manager [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Took 2.13 seconds to destroy the instance on the hypervisor. [ 933.645420] env[62552]: DEBUG oslo.service.loopingcall [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.645420] env[62552]: DEBUG nova.compute.manager [-] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 933.645420] env[62552]: DEBUG nova.network.neutron [-] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 933.687547] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239747, 'name': CloneVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.867954] env[62552]: DEBUG oslo_vmware.api [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239748, 'name': PowerOnVM_Task, 'duration_secs': 0.611522} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.868389] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.868797] env[62552]: INFO nova.compute.manager [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Took 11.49 seconds to spawn the instance on the hypervisor. [ 933.869061] env[62552]: DEBUG nova.compute.manager [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 933.870141] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39fdf336-3ac9-48e2-8070-dce866985c5b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.963474] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.973s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.966940] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.753s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.966940] env[62552]: DEBUG nova.objects.instance [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lazy-loading 'resources' on Instance uuid 4083ee43-ecea-4ea5-8923-42b348893824 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.989574] env[62552]: INFO nova.scheduler.client.report [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Deleted allocations for instance 0080a6b3-cac0-44f3-a037-3f94f1daa275 [ 934.008440] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f02357-069e-41f9-a3a0-08fcd93f7f82 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.014792] env[62552]: DEBUG nova.compute.manager [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 934.034351] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.036811] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e7ab8e8-f015-4114-ab5e-5ca1f680d5ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.043724] env[62552]: DEBUG oslo_vmware.api [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 934.043724] env[62552]: value = "task-1239751" [ 934.043724] env[62552]: _type = "Task" [ 934.043724] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.046011] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.046284] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.046471] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.046666] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.046835] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.047008] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.047242] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.047436] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.047625] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.047801] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.047994] env[62552]: DEBUG nova.virt.hardware [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.049787] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274c8f87-546c-41c8-b73f-ef37ef539881 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.065070] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9689e030-7bcf-464e-be29-c02a8bdcbc08 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.069209] env[62552]: DEBUG oslo_vmware.api [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239751, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.185890] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239747, 'name': CloneVM_Task, 'duration_secs': 1.61922} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.186236] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Created linked-clone VM from snapshot [ 934.187037] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3476204-b0ef-463c-bca0-a76aacbc3dd8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.194671] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Uploading image 0afec343-d526-4ba2-a380-834c5153de71 {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 934.215926] env[62552]: DEBUG nova.compute.manager [req-135e06d0-136c-4a9c-8d67-e241c0f903ea req-a1ee13f0-206c-4749-86d3-5529e5c7d7c4 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Received event network-vif-deleted-f749d931-7a86-41ca-a551-4f8d2013603a {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 934.216245] env[62552]: INFO nova.compute.manager [req-135e06d0-136c-4a9c-8d67-e241c0f903ea req-a1ee13f0-206c-4749-86d3-5529e5c7d7c4 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Neutron deleted interface f749d931-7a86-41ca-a551-4f8d2013603a; detaching it from the instance and deleting it from the info cache [ 934.216497] env[62552]: DEBUG nova.network.neutron [req-135e06d0-136c-4a9c-8d67-e241c0f903ea req-a1ee13f0-206c-4749-86d3-5529e5c7d7c4 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.219556] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 934.219556] env[62552]: value = "vm-267477" [ 934.219556] env[62552]: _type = "VirtualMachine" [ 934.219556] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 934.220585] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-095d5dfe-4136-4b8a-a99e-48e97b7cc331 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.228724] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lease: (returnval){ [ 934.228724] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52751d39-e8ac-d73c-f277-2f5809601048" [ 934.228724] env[62552]: _type = "HttpNfcLease" [ 934.228724] env[62552]: } obtained for exporting VM: (result){ [ 934.228724] env[62552]: value = "vm-267477" [ 934.228724] env[62552]: _type = "VirtualMachine" [ 934.228724] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 934.229013] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the lease: (returnval){ [ 934.229013] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52751d39-e8ac-d73c-f277-2f5809601048" [ 934.229013] env[62552]: _type = "HttpNfcLease" [ 934.229013] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 934.236403] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 934.236403] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52751d39-e8ac-d73c-f277-2f5809601048" [ 934.236403] env[62552]: _type = "HttpNfcLease" [ 934.236403] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 934.327215] env[62552]: DEBUG nova.compute.manager [req-726b179b-0d92-4549-9f64-a8cf3a2b07a6 req-e7839643-3e1a-4e8b-b041-87d55c67a53d service nova] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Received event network-vif-plugged-8018d47b-da5e-40c6-b217-a8b685b44983 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 934.327777] env[62552]: DEBUG oslo_concurrency.lockutils [req-726b179b-0d92-4549-9f64-a8cf3a2b07a6 req-e7839643-3e1a-4e8b-b041-87d55c67a53d service nova] Acquiring lock "c5ed17ee-7421-49d7-975f-1515de1e4b01-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.327777] env[62552]: DEBUG oslo_concurrency.lockutils [req-726b179b-0d92-4549-9f64-a8cf3a2b07a6 req-e7839643-3e1a-4e8b-b041-87d55c67a53d service nova] Lock "c5ed17ee-7421-49d7-975f-1515de1e4b01-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.327965] env[62552]: DEBUG oslo_concurrency.lockutils [req-726b179b-0d92-4549-9f64-a8cf3a2b07a6 req-e7839643-3e1a-4e8b-b041-87d55c67a53d service nova] Lock "c5ed17ee-7421-49d7-975f-1515de1e4b01-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.328246] env[62552]: DEBUG nova.compute.manager [req-726b179b-0d92-4549-9f64-a8cf3a2b07a6 req-e7839643-3e1a-4e8b-b041-87d55c67a53d service nova] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] No waiting events found dispatching network-vif-plugged-8018d47b-da5e-40c6-b217-a8b685b44983 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 934.328522] env[62552]: WARNING nova.compute.manager [req-726b179b-0d92-4549-9f64-a8cf3a2b07a6 req-e7839643-3e1a-4e8b-b041-87d55c67a53d service nova] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Received unexpected event network-vif-plugged-8018d47b-da5e-40c6-b217-a8b685b44983 for instance with vm_state building and task_state spawning. [ 934.391703] env[62552]: INFO nova.compute.manager [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Took 46.90 seconds to build instance. [ 934.426603] env[62552]: DEBUG nova.network.neutron [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Successfully updated port: 8018d47b-da5e-40c6-b217-a8b685b44983 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.504218] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4dd2cd18-69b4-4239-ad16-8d42125ca5b8 tempest-ServerMetadataTestJSON-1887827974 tempest-ServerMetadataTestJSON-1887827974-project-member] Lock "0080a6b3-cac0-44f3-a037-3f94f1daa275" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.250s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.558532] env[62552]: DEBUG oslo_vmware.api [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239751, 'name': PowerOffVM_Task, 'duration_secs': 0.354713} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.560877] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.566228] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Reconfiguring VM instance instance-0000004a to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 934.566706] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a8807b6-8c83-4051-b12b-b72102fcaca7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.586570] env[62552]: DEBUG oslo_vmware.api [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 934.586570] env[62552]: value = "task-1239754" [ 934.586570] env[62552]: _type = "Task" [ 934.586570] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.595939] env[62552]: DEBUG oslo_vmware.api [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239754, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.645929] env[62552]: DEBUG nova.network.neutron [-] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.721106] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bced3db5-5859-4633-8d37-822e681855ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.724150] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e426a24-80d5-44b8-9851-df985a354f04 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.734349] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f07368f-227f-4b0a-866b-a3648b103bbc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.741548] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585686e4-273a-4e4e-8655-76cc546dda10 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.756019] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 934.756019] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52751d39-e8ac-d73c-f277-2f5809601048" [ 934.756019] env[62552]: _type = "HttpNfcLease" [ 934.756019] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 934.779711] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 934.779711] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52751d39-e8ac-d73c-f277-2f5809601048" [ 934.779711] env[62552]: _type = "HttpNfcLease" [ 934.779711] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 934.791087] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4046b895-f8f6-406e-a9ea-ff6a13067d5c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.795236] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b96f82d-6aec-4d26-9a84-365d25f78251 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.796710] env[62552]: DEBUG nova.compute.manager [req-135e06d0-136c-4a9c-8d67-e241c0f903ea req-a1ee13f0-206c-4749-86d3-5529e5c7d7c4 service nova] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Detach interface failed, port_id=f749d931-7a86-41ca-a551-4f8d2013603a, reason: Instance 22357d4e-9771-477c-9fc3-fe3d76f6e902 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 934.805220] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c616820-ced3-4812-9483-69699dceeab4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.808836] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8bb0-33ea-3a8b-7305-197d7f274bbd/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 934.809054] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8bb0-33ea-3a8b-7305-197d7f274bbd/disk-0.vmdk for reading. {{(pid=62552) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 934.876137] env[62552]: DEBUG nova.compute.provider_tree [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.894306] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eeb1ad78-eecc-468e-85ac-17bf5e710c57 tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.961s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.909506] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8e324fab-9084-4323-97dc-0f6dd1de3384 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.929357] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "refresh_cache-c5ed17ee-7421-49d7-975f-1515de1e4b01" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.929556] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "refresh_cache-c5ed17ee-7421-49d7-975f-1515de1e4b01" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.929679] env[62552]: DEBUG nova.network.neutron [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.099381] env[62552]: DEBUG oslo_vmware.api [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239754, 'name': ReconfigVM_Task, 'duration_secs': 0.420631} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.099772] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Reconfigured VM instance instance-0000004a to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 935.100058] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.100329] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3fd56ea-2de4-4ef7-946b-77843917ba20 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.106810] env[62552]: DEBUG oslo_vmware.api [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 935.106810] env[62552]: value = "task-1239755" [ 935.106810] env[62552]: _type = "Task" [ 935.106810] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.118095] env[62552]: DEBUG oslo_vmware.api [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239755, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.150868] env[62552]: INFO nova.compute.manager [-] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Took 1.51 seconds to deallocate network for instance. [ 935.380299] env[62552]: DEBUG nova.scheduler.client.report [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 935.475708] env[62552]: DEBUG nova.network.neutron [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.620490] env[62552]: DEBUG oslo_vmware.api [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239755, 'name': PowerOnVM_Task, 'duration_secs': 0.460741} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.620746] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.620940] env[62552]: DEBUG nova.compute.manager [None req-80018df1-d460-40b9-a615-beacf5cc5b28 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 935.622053] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb22df5-233f-475b-911e-61b69b739358 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.657409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.662796] env[62552]: DEBUG nova.network.neutron [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Updating instance_info_cache with network_info: [{"id": "8018d47b-da5e-40c6-b217-a8b685b44983", "address": "fa:16:3e:3e:a4:ac", "network": {"id": "6ad33460-d660-442d-a742-61dcc66532b5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1461793279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f13a11803d1b44178f053cc0a4affedd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8018d47b-da", "ovs_interfaceid": "8018d47b-da5e-40c6-b217-a8b685b44983", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.886398] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.920s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.890815] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.325s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.891254] env[62552]: DEBUG nova.objects.instance [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lazy-loading 'resources' on Instance uuid 34736dd0-e617-475e-baa2-cb372db1afb2 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.916938] env[62552]: INFO nova.scheduler.client.report [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Deleted allocations for instance 4083ee43-ecea-4ea5-8923-42b348893824 [ 936.166085] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "refresh_cache-c5ed17ee-7421-49d7-975f-1515de1e4b01" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.168950] env[62552]: DEBUG nova.compute.manager [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Instance network_info: |[{"id": "8018d47b-da5e-40c6-b217-a8b685b44983", "address": "fa:16:3e:3e:a4:ac", "network": {"id": "6ad33460-d660-442d-a742-61dcc66532b5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1461793279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f13a11803d1b44178f053cc0a4affedd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8018d47b-da", "ovs_interfaceid": "8018d47b-da5e-40c6-b217-a8b685b44983", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 936.168950] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:a4:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8018d47b-da5e-40c6-b217-a8b685b44983', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.184429] env[62552]: DEBUG oslo.service.loopingcall [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.188022] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 936.188022] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-889e2539-4454-4f0e-ba95-942e1921b2bb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.216035] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.216035] env[62552]: value = "task-1239756" [ 936.216035] env[62552]: _type = "Task" [ 936.216035] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.229221] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239756, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.360124] env[62552]: DEBUG nova.compute.manager [req-d9d477f7-8571-427b-8dfb-f03b1fde54eb req-256c783b-d41a-4a43-a432-5b4c714d00fc service nova] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Received event network-changed-8018d47b-da5e-40c6-b217-a8b685b44983 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 936.360329] env[62552]: DEBUG nova.compute.manager [req-d9d477f7-8571-427b-8dfb-f03b1fde54eb req-256c783b-d41a-4a43-a432-5b4c714d00fc service nova] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Refreshing instance network info cache due to event network-changed-8018d47b-da5e-40c6-b217-a8b685b44983. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 936.360634] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9d477f7-8571-427b-8dfb-f03b1fde54eb req-256c783b-d41a-4a43-a432-5b4c714d00fc service nova] Acquiring lock "refresh_cache-c5ed17ee-7421-49d7-975f-1515de1e4b01" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.365561] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9d477f7-8571-427b-8dfb-f03b1fde54eb req-256c783b-d41a-4a43-a432-5b4c714d00fc service nova] Acquired lock "refresh_cache-c5ed17ee-7421-49d7-975f-1515de1e4b01" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.365561] env[62552]: DEBUG nova.network.neutron [req-d9d477f7-8571-427b-8dfb-f03b1fde54eb req-256c783b-d41a-4a43-a432-5b4c714d00fc service nova] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Refreshing network info cache for port 8018d47b-da5e-40c6-b217-a8b685b44983 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 936.434827] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2269efb2-feca-47e5-8503-fee85981cb39 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "4083ee43-ecea-4ea5-8923-42b348893824" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.024s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.469784] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 936.473043] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267476', 'volume_id': '76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'name': 'volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0', 'attached_at': '', 'detached_at': '', 'volume_id': '76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'serial': '76baa0b3-5dd2-4828-8f24-0aa4063096c8'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 936.473800] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67967771-b70c-498c-8005-a07fa92a32da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.502783] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc09af7-8545-4a16-afb6-0e8322c0ad66 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.533772] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8/volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.537138] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16338998-fa37-4f03-98cc-b4c5ad3383f4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.558894] env[62552]: DEBUG oslo_vmware.api [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 936.558894] env[62552]: value = "task-1239757" [ 936.558894] env[62552]: _type = "Task" [ 936.558894] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.578576] env[62552]: DEBUG oslo_vmware.api [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239757, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.726910] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239756, 'name': CreateVM_Task, 'duration_secs': 0.441545} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.729916] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 936.731580] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.731756] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.732132] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.732423] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ea083b4-b2b4-4784-b3e4-8188539d288e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.737874] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 936.737874] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5265f910-ebc1-7450-9fb3-4e19dfda083e" [ 936.737874] env[62552]: _type = "Task" [ 936.737874] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.743083] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9b4b67-e42a-4b21-a9a4-de007a969bff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.751940] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5265f910-ebc1-7450-9fb3-4e19dfda083e, 'name': SearchDatastore_Task, 'duration_secs': 0.011649} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.753908] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.754179] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.754423] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.754578] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.754759] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.755070] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-babc6667-cfb8-4dc9-b80b-99db1d9a71b3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.757573] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed0d91d-b4cd-49ca-90c2-92ffe07722e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.794886] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cf1205-3fb2-404c-aefe-45e1a6969fef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.798114] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.798320] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.799820] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93085322-342c-450f-8df0-dabdfb97240b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.807221] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 936.807221] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5226e659-78d3-34b3-178a-2cdb5cd85814" [ 936.807221] env[62552]: _type = "Task" [ 936.807221] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.811723] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92158cb-a7b9-491f-8699-5d69caa40120 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.823348] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5226e659-78d3-34b3-178a-2cdb5cd85814, 'name': SearchDatastore_Task, 'duration_secs': 0.013857} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.831355] env[62552]: DEBUG nova.compute.provider_tree [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.832647] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b16eabd-8913-4e3b-af7f-14d8a094c2a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.843163] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 936.843163] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ebc70e-1b96-35d9-a08b-f59ec6031c67" [ 936.843163] env[62552]: _type = "Task" [ 936.843163] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.852374] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ebc70e-1b96-35d9-a08b-f59ec6031c67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.069304] env[62552]: DEBUG oslo_vmware.api [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239757, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.079460] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.079742] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.079998] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.080197] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.080344] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.082663] env[62552]: INFO nova.compute.manager [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Terminating instance [ 937.189576] env[62552]: DEBUG nova.network.neutron [req-d9d477f7-8571-427b-8dfb-f03b1fde54eb req-256c783b-d41a-4a43-a432-5b4c714d00fc service nova] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Updated VIF entry in instance network info cache for port 8018d47b-da5e-40c6-b217-a8b685b44983. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 937.189970] env[62552]: DEBUG nova.network.neutron [req-d9d477f7-8571-427b-8dfb-f03b1fde54eb req-256c783b-d41a-4a43-a432-5b4c714d00fc service nova] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Updating instance_info_cache with network_info: [{"id": "8018d47b-da5e-40c6-b217-a8b685b44983", "address": "fa:16:3e:3e:a4:ac", "network": {"id": "6ad33460-d660-442d-a742-61dcc66532b5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1461793279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f13a11803d1b44178f053cc0a4affedd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8018d47b-da", "ovs_interfaceid": "8018d47b-da5e-40c6-b217-a8b685b44983", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.337587] env[62552]: DEBUG nova.scheduler.client.report [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 937.353366] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ebc70e-1b96-35d9-a08b-f59ec6031c67, 'name': SearchDatastore_Task, 'duration_secs': 0.01386} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.354426] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.354736] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] c5ed17ee-7421-49d7-975f-1515de1e4b01/c5ed17ee-7421-49d7-975f-1515de1e4b01.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.355198] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ca4050d-06ef-4e0c-8345-bc6ec42e5627 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.362447] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 937.362447] env[62552]: value = "task-1239758" [ 937.362447] env[62552]: _type = "Task" [ 937.362447] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.371054] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239758, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.422922] env[62552]: DEBUG oslo_concurrency.lockutils [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.423301] env[62552]: DEBUG oslo_concurrency.lockutils [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.423611] env[62552]: DEBUG oslo_concurrency.lockutils [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.423846] env[62552]: DEBUG oslo_concurrency.lockutils [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.424389] env[62552]: DEBUG oslo_concurrency.lockutils [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.428093] env[62552]: INFO nova.compute.manager [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Terminating instance [ 937.572872] env[62552]: DEBUG oslo_vmware.api [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239757, 'name': ReconfigVM_Task, 'duration_secs': 0.639786} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.573259] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Reconfigured VM instance instance-00000049 to attach disk [datastore2] volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8/volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.578360] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40bf4efc-dfb5-4229-9278-e30760aae74e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.589290] env[62552]: DEBUG nova.compute.manager [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 937.589522] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.590388] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcab5e21-a18c-4481-967c-3d6ffb35362c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.598597] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.599888] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a58df8a5-2baf-49bc-875d-0f59236579cb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.601664] env[62552]: DEBUG oslo_vmware.api [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 937.601664] env[62552]: value = "task-1239759" [ 937.601664] env[62552]: _type = "Task" [ 937.601664] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.606579] env[62552]: DEBUG oslo_vmware.api [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 937.606579] env[62552]: value = "task-1239760" [ 937.606579] env[62552]: _type = "Task" [ 937.606579] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.609899] env[62552]: DEBUG oslo_vmware.api [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239759, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.618338] env[62552]: DEBUG oslo_vmware.api [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239760, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.693747] env[62552]: DEBUG oslo_concurrency.lockutils [req-d9d477f7-8571-427b-8dfb-f03b1fde54eb req-256c783b-d41a-4a43-a432-5b4c714d00fc service nova] Releasing lock "refresh_cache-c5ed17ee-7421-49d7-975f-1515de1e4b01" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.842688] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.952s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.845872] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.019s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.846292] env[62552]: DEBUG nova.objects.instance [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lazy-loading 'resources' on Instance uuid 714fb65d-9f80-4a81-a637-3e4398405d9b {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.877297] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239758, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.878689] env[62552]: INFO nova.scheduler.client.report [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleted allocations for instance 34736dd0-e617-475e-baa2-cb372db1afb2 [ 937.935216] env[62552]: DEBUG nova.compute.manager [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 937.935216] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.935499] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e352ea-efc9-4e18-8a8e-6ef7e6f6869f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.946798] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.947251] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-106dbe1e-08f6-49d7-875c-b0829e0abed6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.957702] env[62552]: DEBUG oslo_vmware.api [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 937.957702] env[62552]: value = "task-1239761" [ 937.957702] env[62552]: _type = "Task" [ 937.957702] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.968762] env[62552]: DEBUG oslo_vmware.api [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239761, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.112964] env[62552]: DEBUG oslo_vmware.api [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239759, 'name': ReconfigVM_Task, 'duration_secs': 0.169817} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.117531] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267476', 'volume_id': '76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'name': 'volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0', 'attached_at': '', 'detached_at': '', 'volume_id': '76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'serial': '76baa0b3-5dd2-4828-8f24-0aa4063096c8'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 938.124953] env[62552]: DEBUG oslo_vmware.api [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239760, 'name': PowerOffVM_Task, 'duration_secs': 0.453439} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.125409] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.125664] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.125978] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-655475a1-c375-4f49-a5f3-239c6ee00f07 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.198748] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.199050] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.199358] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Deleting the datastore file [datastore1] ae517e5f-0cd1-457c-99a6-6cd46fdd6b06 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.199703] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-719145ff-f08c-4985-a1b6-40c1109cd116 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.205939] env[62552]: DEBUG oslo_vmware.api [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 938.205939] env[62552]: value = "task-1239763" [ 938.205939] env[62552]: _type = "Task" [ 938.205939] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.214893] env[62552]: DEBUG oslo_vmware.api [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239763, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.373133] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239758, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641288} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.376059] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] c5ed17ee-7421-49d7-975f-1515de1e4b01/c5ed17ee-7421-49d7-975f-1515de1e4b01.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.376335] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.376837] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84f92eff-eaf4-46d0-adb3-c35294842f6a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.383189] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 938.383189] env[62552]: value = "task-1239764" [ 938.383189] env[62552]: _type = "Task" [ 938.383189] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.393308] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e8ea12fc-f5ee-42ef-bc0c-ba210d53b0dc tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "34736dd0-e617-475e-baa2-cb372db1afb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.971s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.397730] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239764, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.469975] env[62552]: DEBUG oslo_vmware.api [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239761, 'name': PowerOffVM_Task, 'duration_secs': 0.213549} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.470285] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.470461] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.470751] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22298a51-cddb-44bf-b995-3bae52a607e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.579089] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.579596] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.580199] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Deleting the datastore file [datastore2] d8891025-5bdf-4dc9-a2b0-c86f94582ac6 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.580297] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-76b0a1f3-52a0-41a3-8f52-b253d4b50a9c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.591449] env[62552]: DEBUG oslo_vmware.api [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for the task: (returnval){ [ 938.591449] env[62552]: value = "task-1239766" [ 938.591449] env[62552]: _type = "Task" [ 938.591449] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.600051] env[62552]: DEBUG oslo_vmware.api [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239766, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.636412] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb63d17-6c23-4d0d-b116-223419a57b70 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.650136] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99731c82-a03e-414a-a6b7-246bcd56500f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.684799] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5457c55-318e-4342-a1b3-5ef5051d648e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.693860] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2440ffa-3720-48d5-990f-69233af2a1ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.708567] env[62552]: DEBUG nova.compute.provider_tree [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.718479] env[62552]: DEBUG oslo_vmware.api [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239763, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369885} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.719369] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.720027] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.720027] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.720027] env[62552]: INFO nova.compute.manager [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Took 1.13 seconds to destroy the instance on the hypervisor. [ 938.720202] env[62552]: DEBUG oslo.service.loopingcall [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.720348] env[62552]: DEBUG nova.compute.manager [-] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 938.720443] env[62552]: DEBUG nova.network.neutron [-] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 938.893667] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239764, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.149723} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.893985] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 938.894758] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ffa0d7-e947-4693-b586-ded44b0e72d3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.918840] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] c5ed17ee-7421-49d7-975f-1515de1e4b01/c5ed17ee-7421-49d7-975f-1515de1e4b01.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 938.919643] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69f222f8-dbb1-4b0e-85bf-004fe68c1262 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.941457] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 938.941457] env[62552]: value = "task-1239767" [ 938.941457] env[62552]: _type = "Task" [ 938.941457] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.949386] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239767, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.102953] env[62552]: DEBUG oslo_vmware.api [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Task: {'id': task-1239766, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.32289} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.103800] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.103998] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.104314] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.104519] env[62552]: INFO nova.compute.manager [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Took 1.17 seconds to destroy the instance on the hypervisor. [ 939.104894] env[62552]: DEBUG oslo.service.loopingcall [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.105224] env[62552]: DEBUG nova.compute.manager [-] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 939.105385] env[62552]: DEBUG nova.network.neutron [-] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 939.189091] env[62552]: DEBUG nova.objects.instance [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.214672] env[62552]: DEBUG nova.scheduler.client.report [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 939.242741] env[62552]: DEBUG nova.compute.manager [req-c1397022-441a-47fa-a888-d0cefab977fc req-4c916c0c-11b3-417b-85d1-bb681ba8ae88 service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Received event network-vif-deleted-cd1b0e6a-5546-47af-84df-a5e05587638d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 939.243616] env[62552]: INFO nova.compute.manager [req-c1397022-441a-47fa-a888-d0cefab977fc req-4c916c0c-11b3-417b-85d1-bb681ba8ae88 service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Neutron deleted interface cd1b0e6a-5546-47af-84df-a5e05587638d; detaching it from the instance and deleting it from the info cache [ 939.243848] env[62552]: DEBUG nova.network.neutron [req-c1397022-441a-47fa-a888-d0cefab977fc req-4c916c0c-11b3-417b-85d1-bb681ba8ae88 service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.450673] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239767, 'name': ReconfigVM_Task, 'duration_secs': 0.387745} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.451187] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Reconfigured VM instance instance-00000053 to attach disk [datastore2] c5ed17ee-7421-49d7-975f-1515de1e4b01/c5ed17ee-7421-49d7-975f-1515de1e4b01.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.451738] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5529ae1a-2d7d-413d-a0f5-c2dd30f59185 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.457557] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 939.457557] env[62552]: value = "task-1239768" [ 939.457557] env[62552]: _type = "Task" [ 939.457557] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.467425] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239768, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.569103] env[62552]: DEBUG nova.network.neutron [-] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.698298] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a83e885f-01d3-4846-905b-982dd6cb4cad tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.889s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.724674] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.726616] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.787s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.728294] env[62552]: INFO nova.compute.claims [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.748236] env[62552]: INFO nova.scheduler.client.report [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Deleted allocations for instance 714fb65d-9f80-4a81-a637-3e4398405d9b [ 939.753138] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79ff73f7-1d5e-4a16-9c74-06c58318228e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.762757] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3274c6c0-d1da-4957-8855-dccda5149092 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.802439] env[62552]: DEBUG nova.compute.manager [req-c1397022-441a-47fa-a888-d0cefab977fc req-4c916c0c-11b3-417b-85d1-bb681ba8ae88 service nova] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Detach interface failed, port_id=cd1b0e6a-5546-47af-84df-a5e05587638d, reason: Instance ae517e5f-0cd1-457c-99a6-6cd46fdd6b06 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 939.860159] env[62552]: DEBUG nova.network.neutron [-] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.970710] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239768, 'name': Rename_Task, 'duration_secs': 0.185196} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.971100] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.971591] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c7bc544-ffcb-4279-8e4b-2a23c91a8592 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.978660] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 939.978660] env[62552]: value = "task-1239769" [ 939.978660] env[62552]: _type = "Task" [ 939.978660] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.988485] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239769, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.055534] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "79166f2c-f864-4d8a-b1dc-e176710400cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.055785] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.072416] env[62552]: INFO nova.compute.manager [-] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Took 1.35 seconds to deallocate network for instance. [ 940.259024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-37ef52ac-5a75-4c2f-a92e-5a842f8de317 tempest-ServerRescueTestJSON-1699739245 tempest-ServerRescueTestJSON-1699739245-project-member] Lock "714fb65d-9f80-4a81-a637-3e4398405d9b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.541s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.279881] env[62552]: DEBUG oslo_concurrency.lockutils [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.280187] env[62552]: DEBUG oslo_concurrency.lockutils [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.280400] env[62552]: DEBUG nova.compute.manager [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 940.281700] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b80f34-62d1-4430-8a37-3c88931c62ac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.288929] env[62552]: DEBUG nova.compute.manager [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62552) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 940.289571] env[62552]: DEBUG nova.objects.instance [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.363045] env[62552]: INFO nova.compute.manager [-] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Took 1.26 seconds to deallocate network for instance. [ 940.492049] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239769, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.558528] env[62552]: DEBUG nova.compute.manager [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 940.579186] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.869354] env[62552]: DEBUG oslo_concurrency.lockutils [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.989683] env[62552]: DEBUG oslo_vmware.api [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239769, 'name': PowerOnVM_Task, 'duration_secs': 0.594528} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.992566] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.993103] env[62552]: INFO nova.compute.manager [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Took 6.98 seconds to spawn the instance on the hypervisor. [ 940.993103] env[62552]: DEBUG nova.compute.manager [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 940.993938] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32931100-2e53-4c62-be8a-61535731d04c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.021978] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2fa5a6-72ca-4863-bba9-73a5cd122b7b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.028976] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0880581b-d9b1-4cfb-971f-5f931affd111 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.061989] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88dd5e83-432c-47b3-b5bd-408639b96134 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.074189] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8b4e07-c669-47ca-abee-46def0b0ace1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.088162] env[62552]: DEBUG nova.compute.provider_tree [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.089990] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.277225] env[62552]: DEBUG nova.compute.manager [req-c4841ded-6b0c-439b-b8b9-ff7a232c348d req-dbef912e-7e25-4558-803c-4363174382c4 service nova] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Received event network-vif-deleted-aee2bcfd-d515-46fa-8574-2044d313840b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 941.299739] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 941.300069] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e3ab20f-0ba5-4e21-8ea6-9cbfc051ec2f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.308233] env[62552]: DEBUG oslo_vmware.api [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 941.308233] env[62552]: value = "task-1239770" [ 941.308233] env[62552]: _type = "Task" [ 941.308233] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.318435] env[62552]: DEBUG oslo_vmware.api [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239770, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.515596] env[62552]: INFO nova.compute.manager [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Took 15.01 seconds to build instance. [ 941.595151] env[62552]: DEBUG nova.scheduler.client.report [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 941.818970] env[62552]: DEBUG oslo_vmware.api [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239770, 'name': PowerOffVM_Task, 'duration_secs': 0.226378} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.819284] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.819490] env[62552]: DEBUG nova.compute.manager [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 941.820350] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e35150-3510-412f-9151-79db6bfa0844 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.017647] env[62552]: DEBUG oslo_concurrency.lockutils [None req-81a5ed75-e6b5-47fd-a6e1-abcbf14f1c17 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "c5ed17ee-7421-49d7-975f-1515de1e4b01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.523s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.101730] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.102276] env[62552]: DEBUG nova.compute.manager [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 942.104862] env[62552]: DEBUG oslo_concurrency.lockutils [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.664s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.105151] env[62552]: DEBUG nova.objects.instance [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'resources' on Instance uuid e31e3077-04e1-4adb-a0cc-44cf84a89eda {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.334932] env[62552]: DEBUG oslo_concurrency.lockutils [None req-15468763-2e2a-48e0-9b39-5237b7d21803 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.055s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.611967] env[62552]: DEBUG nova.compute.utils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.613652] env[62552]: DEBUG nova.compute.manager [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 942.613825] env[62552]: DEBUG nova.network.neutron [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 942.666326] env[62552]: DEBUG nova.policy [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c732a4a15f9b4b16b241c09f4ba373e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3daa5253a0cc4538903b7cabc2f98fa2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.779062] env[62552]: DEBUG nova.objects.instance [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.867268] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4850abe9-4aa0-4484-9bc0-3f97ae016ddf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.877889] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79794e12-99ee-465c-89fd-a8c1ed5a88d7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.912603] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe31066b-95fc-4401-bbbc-e1670be1b7bd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.920506] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0cd156-754a-4092-9ba6-c41c0def39e1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.936415] env[62552]: DEBUG nova.compute.provider_tree [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.956224] env[62552]: DEBUG nova.network.neutron [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Successfully created port: 04e39cd5-39ae-4671-b715-e300dec6021d {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.117408] env[62552]: DEBUG nova.compute.manager [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 943.288863] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.289088] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.289306] env[62552]: DEBUG nova.network.neutron [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.289566] env[62552]: DEBUG nova.objects.instance [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'info_cache' on Instance uuid a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.439571] env[62552]: DEBUG nova.scheduler.client.report [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 943.647153] env[62552]: DEBUG nova.compute.manager [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 943.648108] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a5c585-eede-4b53-973b-842cd6e395e1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.792874] env[62552]: DEBUG nova.objects.base [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 943.944917] env[62552]: DEBUG oslo_concurrency.lockutils [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.840s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.947493] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.290s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.947957] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.950352] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.371s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.950704] env[62552]: DEBUG nova.objects.instance [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lazy-loading 'resources' on Instance uuid ae517e5f-0cd1-457c-99a6-6cd46fdd6b06 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.967513] env[62552]: INFO nova.scheduler.client.report [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleted allocations for instance e31e3077-04e1-4adb-a0cc-44cf84a89eda [ 943.969421] env[62552]: INFO nova.scheduler.client.report [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted allocations for instance 22357d4e-9771-477c-9fc3-fe3d76f6e902 [ 944.127176] env[62552]: DEBUG nova.compute.manager [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 944.151242] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.151564] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.151740] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.151929] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.152091] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.152246] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.152478] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.152651] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.152823] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.152988] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.153185] env[62552]: DEBUG nova.virt.hardware [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.154082] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf2f9f6-fb4f-4de5-92d9-318554ac4754 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.158727] env[62552]: INFO nova.compute.manager [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] instance snapshotting [ 944.163490] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6884e4b4-faf6-4190-8926-0ef181178de5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.167942] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c00284-aecf-4bb2-be2a-9e1538b2ac29 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.193132] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bd4171-c75b-49ec-a1fe-cf06042f395a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.389304] env[62552]: DEBUG nova.compute.manager [req-e9e12b41-26ab-4852-ae19-da548d2523eb req-b2ef4d53-3291-4723-a0b9-e6a702dc180d service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Received event network-vif-plugged-04e39cd5-39ae-4671-b715-e300dec6021d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 944.389478] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9e12b41-26ab-4852-ae19-da548d2523eb req-b2ef4d53-3291-4723-a0b9-e6a702dc180d service nova] Acquiring lock "04186fde-a9ed-415a-bd40-312f0347fcc7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.389697] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9e12b41-26ab-4852-ae19-da548d2523eb req-b2ef4d53-3291-4723-a0b9-e6a702dc180d service nova] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.389871] env[62552]: DEBUG oslo_concurrency.lockutils [req-e9e12b41-26ab-4852-ae19-da548d2523eb req-b2ef4d53-3291-4723-a0b9-e6a702dc180d service nova] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.390436] env[62552]: DEBUG nova.compute.manager [req-e9e12b41-26ab-4852-ae19-da548d2523eb req-b2ef4d53-3291-4723-a0b9-e6a702dc180d service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] No waiting events found dispatching network-vif-plugged-04e39cd5-39ae-4671-b715-e300dec6021d {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 944.390720] env[62552]: WARNING nova.compute.manager [req-e9e12b41-26ab-4852-ae19-da548d2523eb req-b2ef4d53-3291-4723-a0b9-e6a702dc180d service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Received unexpected event network-vif-plugged-04e39cd5-39ae-4671-b715-e300dec6021d for instance with vm_state building and task_state spawning. [ 944.475788] env[62552]: DEBUG nova.network.neutron [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Successfully updated port: 04e39cd5-39ae-4671-b715-e300dec6021d {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 944.483511] env[62552]: DEBUG oslo_concurrency.lockutils [None req-28549240-94b8-4ef4-83f9-273ff7dd08e3 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "e31e3077-04e1-4adb-a0cc-44cf84a89eda" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.010s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.484609] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d44e60d4-226c-4a78-b8c5-85fb0b223502 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "22357d4e-9771-477c-9fc3-fe3d76f6e902" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.473s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.531483] env[62552]: DEBUG nova.network.neutron [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updating instance_info_cache with network_info: [{"id": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "address": "fa:16:3e:fc:98:ec", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56e2a627-19", "ovs_interfaceid": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.684690] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d44b251-75c7-4e96-862c-31a8be7f0ef1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.692605] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82197c47-16ac-4641-9581-142ef7f17299 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.727238] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 944.730055] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ff5bb4d3-ff3e-433a-ae93-7fcd1c901825 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.731787] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219d372b-ac9c-4ff9-ab82-19f6168b9138 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.739336] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a2230e-eee4-4e40-9e60-398832c1f094 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.745015] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8bb0-33ea-3a8b-7305-197d7f274bbd/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 944.746900] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4666aedd-c2e3-4976-813c-49d242899e2b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.749489] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 944.749489] env[62552]: value = "task-1239771" [ 944.749489] env[62552]: _type = "Task" [ 944.749489] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.762874] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8bb0-33ea-3a8b-7305-197d7f274bbd/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 944.763631] env[62552]: ERROR oslo_vmware.rw_handles [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8bb0-33ea-3a8b-7305-197d7f274bbd/disk-0.vmdk due to incomplete transfer. [ 944.763631] env[62552]: DEBUG nova.compute.provider_tree [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.764996] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3a83525c-7904-4d58-b78c-7ac85a8d84ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.770264] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239771, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.776500] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8bb0-33ea-3a8b-7305-197d7f274bbd/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 944.776712] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Uploaded image 0afec343-d526-4ba2-a380-834c5153de71 to the Glance image server {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 944.778956] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 944.779263] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6a05b2f1-11cd-4261-ac22-9f0e491e10c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.785779] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 944.785779] env[62552]: value = "task-1239772" [ 944.785779] env[62552]: _type = "Task" [ 944.785779] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.794844] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239772, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.984707] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.984884] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.985063] env[62552]: DEBUG nova.network.neutron [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 945.035200] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Releasing lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.260256] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239771, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.271218] env[62552]: DEBUG nova.scheduler.client.report [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 945.298253] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239772, 'name': Destroy_Task, 'duration_secs': 0.39675} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.298253] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Destroyed the VM [ 945.298253] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 945.298253] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1e436b81-8792-40ae-a134-d646930eecc4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.310027] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 945.310027] env[62552]: value = "task-1239773" [ 945.310027] env[62552]: _type = "Task" [ 945.310027] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.324532] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239773, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.548104] env[62552]: DEBUG nova.network.neutron [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 945.726910] env[62552]: DEBUG nova.network.neutron [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance_info_cache with network_info: [{"id": "04e39cd5-39ae-4671-b715-e300dec6021d", "address": "fa:16:3e:6e:65:88", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04e39cd5-39", "ovs_interfaceid": "04e39cd5-39ae-4671-b715-e300dec6021d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.761195] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239771, 'name': CreateSnapshot_Task, 'duration_secs': 0.769184} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.761532] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 945.762345] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ebf4e2-07de-4c7c-b88f-9bdcbcea12c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.776926] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.826s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.780791] env[62552]: DEBUG oslo_concurrency.lockutils [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.910s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.780791] env[62552]: DEBUG nova.objects.instance [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lazy-loading 'resources' on Instance uuid d8891025-5bdf-4dc9-a2b0-c86f94582ac6 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.812951] env[62552]: INFO nova.scheduler.client.report [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Deleted allocations for instance ae517e5f-0cd1-457c-99a6-6cd46fdd6b06 [ 945.819415] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239773, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.042190] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.042519] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30fdface-b735-4c21-9b76-fd3931e78334 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.051931] env[62552]: DEBUG oslo_vmware.api [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 946.051931] env[62552]: value = "task-1239774" [ 946.051931] env[62552]: _type = "Task" [ 946.051931] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.058634] env[62552]: DEBUG oslo_vmware.api [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.233193] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.233551] env[62552]: DEBUG nova.compute.manager [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Instance network_info: |[{"id": "04e39cd5-39ae-4671-b715-e300dec6021d", "address": "fa:16:3e:6e:65:88", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04e39cd5-39", "ovs_interfaceid": "04e39cd5-39ae-4671-b715-e300dec6021d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 946.234030] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:65:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e41c97-4d75-4041-ae71-321e7e9d480b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04e39cd5-39ae-4671-b715-e300dec6021d', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.241641] env[62552]: DEBUG oslo.service.loopingcall [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.241861] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.242101] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ae7b3b8-ece4-4128-add9-25756bc719d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.263559] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.263559] env[62552]: value = "task-1239775" [ 946.263559] env[62552]: _type = "Task" [ 946.263559] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.271315] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239775, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.281200] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 946.281516] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7369c770-edcd-4372-ba6b-4e9aab7f0df6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.292618] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 946.292618] env[62552]: value = "task-1239776" [ 946.292618] env[62552]: _type = "Task" [ 946.292618] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.301287] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239776, 'name': CloneVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.320523] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239773, 'name': RemoveSnapshot_Task, 'duration_secs': 0.869584} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.320880] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 946.321100] env[62552]: DEBUG nova.compute.manager [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 946.321916] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a32ccc6-8ac2-43f8-bce0-baf00c000c64 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.329421] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9cb53cbd-202f-4452-8da7-71f1a7ff6d1c tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "ae517e5f-0cd1-457c-99a6-6cd46fdd6b06" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.250s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.423083] env[62552]: DEBUG nova.compute.manager [req-4e7af557-0110-4ef7-a5be-cfcfe2f742c5 req-bb790a02-2553-42d3-a988-abb16359a77f service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Received event network-changed-04e39cd5-39ae-4671-b715-e300dec6021d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 946.423083] env[62552]: DEBUG nova.compute.manager [req-4e7af557-0110-4ef7-a5be-cfcfe2f742c5 req-bb790a02-2553-42d3-a988-abb16359a77f service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Refreshing instance network info cache due to event network-changed-04e39cd5-39ae-4671-b715-e300dec6021d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 946.423083] env[62552]: DEBUG oslo_concurrency.lockutils [req-4e7af557-0110-4ef7-a5be-cfcfe2f742c5 req-bb790a02-2553-42d3-a988-abb16359a77f service nova] Acquiring lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.423083] env[62552]: DEBUG oslo_concurrency.lockutils [req-4e7af557-0110-4ef7-a5be-cfcfe2f742c5 req-bb790a02-2553-42d3-a988-abb16359a77f service nova] Acquired lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.423083] env[62552]: DEBUG nova.network.neutron [req-4e7af557-0110-4ef7-a5be-cfcfe2f742c5 req-bb790a02-2553-42d3-a988-abb16359a77f service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Refreshing network info cache for port 04e39cd5-39ae-4671-b715-e300dec6021d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.510379] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e38aaf-68db-4110-b575-6860df545fad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.519998] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a078ec8-c9e7-4190-9e01-749bbe887d15 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.555217] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17350221-222e-4247-83af-7fb4eaf2c576 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.563147] env[62552]: DEBUG oslo_vmware.api [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239774, 'name': PowerOnVM_Task, 'duration_secs': 0.472146} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.566024] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.566024] env[62552]: DEBUG nova.compute.manager [None req-4cd92f1f-b7ad-4efd-8600-c11e7a210deb tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 946.566600] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337ad258-7438-46bd-909a-425540365b74 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.569963] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1100c69-ec42-48e8-96c6-10ba48b63ad8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.587212] env[62552]: DEBUG nova.compute.provider_tree [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.773824] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239775, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.808510] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239776, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.837942] env[62552]: INFO nova.compute.manager [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Shelve offloading [ 947.092271] env[62552]: DEBUG nova.scheduler.client.report [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 947.269449] env[62552]: DEBUG nova.network.neutron [req-4e7af557-0110-4ef7-a5be-cfcfe2f742c5 req-bb790a02-2553-42d3-a988-abb16359a77f service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updated VIF entry in instance network info cache for port 04e39cd5-39ae-4671-b715-e300dec6021d. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.270349] env[62552]: DEBUG nova.network.neutron [req-4e7af557-0110-4ef7-a5be-cfcfe2f742c5 req-bb790a02-2553-42d3-a988-abb16359a77f service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance_info_cache with network_info: [{"id": "04e39cd5-39ae-4671-b715-e300dec6021d", "address": "fa:16:3e:6e:65:88", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04e39cd5-39", "ovs_interfaceid": "04e39cd5-39ae-4671-b715-e300dec6021d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.276433] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239775, 'name': CreateVM_Task, 'duration_secs': 0.51428} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.276629] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.277262] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.277494] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.278305] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.278305] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de7bb08a-fe71-4f51-9df3-dd7c2fb8983f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.282533] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 947.282533] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fc0619-8293-f96e-ad05-f94b84a0d741" [ 947.282533] env[62552]: _type = "Task" [ 947.282533] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.291026] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fc0619-8293-f96e-ad05-f94b84a0d741, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.300855] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239776, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.342017] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 947.342701] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f01db51-55a5-4c37-9d35-7244df9b5e28 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.349827] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 947.349827] env[62552]: value = "task-1239778" [ 947.349827] env[62552]: _type = "Task" [ 947.349827] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.358496] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239778, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.592060] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.592637] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.599504] env[62552]: DEBUG oslo_concurrency.lockutils [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.820s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.602560] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.513s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.604111] env[62552]: INFO nova.compute.claims [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.617530] env[62552]: INFO nova.scheduler.client.report [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Deleted allocations for instance d8891025-5bdf-4dc9-a2b0-c86f94582ac6 [ 947.772411] env[62552]: DEBUG oslo_concurrency.lockutils [req-4e7af557-0110-4ef7-a5be-cfcfe2f742c5 req-bb790a02-2553-42d3-a988-abb16359a77f service nova] Releasing lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.795078] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fc0619-8293-f96e-ad05-f94b84a0d741, 'name': SearchDatastore_Task, 'duration_secs': 0.008946} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.797960] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.798226] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.798463] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.798615] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.798797] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.799059] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b1629e2-5346-490f-8792-175ef85e87ef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.805663] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239776, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.806921] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.807129] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.807793] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-863d9683-e9e7-4335-9844-97b82d6ca52c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.812281] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 947.812281] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f700f8-cc5e-7973-0aad-fed454e7514e" [ 947.812281] env[62552]: _type = "Task" [ 947.812281] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.819938] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f700f8-cc5e-7973-0aad-fed454e7514e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.824352] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.824572] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.824781] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.824966] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.825151] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.827150] env[62552]: INFO nova.compute.manager [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Terminating instance [ 947.862539] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 947.862814] env[62552]: DEBUG nova.compute.manager [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 947.863692] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9132dfe0-6769-4268-a6f5-6fb89977ef1e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.870414] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.870645] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.870865] env[62552]: DEBUG nova.network.neutron [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.094422] env[62552]: DEBUG nova.compute.manager [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 948.125021] env[62552]: DEBUG oslo_concurrency.lockutils [None req-98489431-e583-4b92-88e7-f5b06fa601c7 tempest-ListImageFiltersTestJSON-1815332097 tempest-ListImageFiltersTestJSON-1815332097-project-member] Lock "d8891025-5bdf-4dc9-a2b0-c86f94582ac6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.702s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.304213] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239776, 'name': CloneVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.324161] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f700f8-cc5e-7973-0aad-fed454e7514e, 'name': SearchDatastore_Task, 'duration_secs': 0.007986} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.325165] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3e2594a-1a72-4baf-a8d7-7e01bcd3c12a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.331328] env[62552]: DEBUG nova.compute.manager [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 948.331689] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 948.332184] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 948.332184] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a7d065-dbcb-99aa-8f6c-65dab4cc14ad" [ 948.332184] env[62552]: _type = "Task" [ 948.332184] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.333213] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61336c51-7195-4199-a6ab-c6f79e20d15c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.344995] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a7d065-dbcb-99aa-8f6c-65dab4cc14ad, 'name': SearchDatastore_Task, 'duration_secs': 0.0089} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.347152] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.347837] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 04186fde-a9ed-415a-bd40-312f0347fcc7/04186fde-a9ed-415a-bd40-312f0347fcc7.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 948.347837] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.348019] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-99f750e8-3abd-4a6b-bc1d-29f52c97add7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.349889] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1886221-04f2-46ea-a94d-402e32913767 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.355583] env[62552]: DEBUG oslo_vmware.api [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 948.355583] env[62552]: value = "task-1239780" [ 948.355583] env[62552]: _type = "Task" [ 948.355583] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.356712] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 948.356712] env[62552]: value = "task-1239779" [ 948.356712] env[62552]: _type = "Task" [ 948.356712] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.366770] env[62552]: DEBUG oslo_vmware.api [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239780, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.369845] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239779, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.631902] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.742937] env[62552]: DEBUG nova.network.neutron [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Updating instance_info_cache with network_info: [{"id": "61514a33-d741-4aae-876d-b0ecbe6ed6e6", "address": "fa:16:3e:b9:cf:bd", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61514a33-d7", "ovs_interfaceid": "61514a33-d741-4aae-876d-b0ecbe6ed6e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.808545] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239776, 'name': CloneVM_Task, 'duration_secs': 2.040191} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.811143] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Created linked-clone VM from snapshot [ 948.812104] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f873ab9-e1bf-4dfd-86bd-dad67fa0c668 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.819511] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Uploading image d2991aba-c054-4e7b-90f5-6d2811506639 {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 948.833499] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be73f18-6abb-470c-b948-9435ed07ec16 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.837017] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 948.837274] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9889c91d-8c32-4c85-8eba-58b1986dd1d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.844267] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968bd5b8-d8fb-4c38-bfdd-1d517fc27a1b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.847475] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 948.847475] env[62552]: value = "task-1239781" [ 948.847475] env[62552]: _type = "Task" [ 948.847475] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.885031] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b71340c-b84a-452f-8fff-6e72bd738f8f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.890302] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239781, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.895774] env[62552]: DEBUG oslo_vmware.api [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239780, 'name': PowerOffVM_Task, 'duration_secs': 0.1837} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.901065] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.901299] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 948.901667] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239779, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458467} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.901971] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9710f0db-8b50-4c02-962b-48ea1ea8323b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.903692] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 04186fde-a9ed-415a-bd40-312f0347fcc7/04186fde-a9ed-415a-bd40-312f0347fcc7.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.903940] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.905249] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e95d7f-42b1-4ff3-a6e2-0bec7606c76f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.909741] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f46f739-4483-4429-8a12-4de6ec9459cb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.925336] env[62552]: DEBUG nova.compute.provider_tree [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.925336] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 948.925336] env[62552]: value = "task-1239783" [ 948.925336] env[62552]: _type = "Task" [ 948.925336] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.932876] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239783, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.975594] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 948.976292] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 948.976292] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleting the datastore file [datastore2] fbbc360d-9cfd-48f1-80b2-26da2c72c002 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.976292] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25445883-6fcb-458d-9cf5-1ff2b37c89b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.983607] env[62552]: DEBUG oslo_vmware.api [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 948.983607] env[62552]: value = "task-1239784" [ 948.983607] env[62552]: _type = "Task" [ 948.983607] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.993140] env[62552]: DEBUG oslo_vmware.api [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239784, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.246364] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.357156] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239781, 'name': Destroy_Task, 'duration_secs': 0.358113} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.357469] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Destroyed the VM [ 949.357668] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 949.357916] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-952ced60-f8db-4d32-b227-fb059793bebc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.364378] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 949.364378] env[62552]: value = "task-1239785" [ 949.364378] env[62552]: _type = "Task" [ 949.364378] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.372586] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239785, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.429016] env[62552]: DEBUG nova.scheduler.client.report [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 949.437994] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239783, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079052} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.437994] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.440310] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b29283-6081-4d9a-9d1b-7d8824276542 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.463071] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 04186fde-a9ed-415a-bd40-312f0347fcc7/04186fde-a9ed-415a-bd40-312f0347fcc7.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.463813] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4a3285e-bcee-4839-82fc-b53acbec7ed7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.484089] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 949.484089] env[62552]: value = "task-1239786" [ 949.484089] env[62552]: _type = "Task" [ 949.484089] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.496239] env[62552]: DEBUG oslo_vmware.api [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239784, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143943} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.499759] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 949.499965] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 949.500168] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 949.500348] env[62552]: INFO nova.compute.manager [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Took 1.17 seconds to destroy the instance on the hypervisor. [ 949.500632] env[62552]: DEBUG oslo.service.loopingcall [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.500866] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239786, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.501095] env[62552]: DEBUG nova.compute.manager [-] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 949.501214] env[62552]: DEBUG nova.network.neutron [-] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 949.654449] env[62552]: DEBUG nova.compute.manager [req-301827d5-93cc-4cb9-96ff-ba8780077be3 req-5088f3ff-90ed-44a8-ab97-12a6ed8a5d6b service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Received event network-vif-unplugged-61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 949.654675] env[62552]: DEBUG oslo_concurrency.lockutils [req-301827d5-93cc-4cb9-96ff-ba8780077be3 req-5088f3ff-90ed-44a8-ab97-12a6ed8a5d6b service nova] Acquiring lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.654888] env[62552]: DEBUG oslo_concurrency.lockutils [req-301827d5-93cc-4cb9-96ff-ba8780077be3 req-5088f3ff-90ed-44a8-ab97-12a6ed8a5d6b service nova] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.655071] env[62552]: DEBUG oslo_concurrency.lockutils [req-301827d5-93cc-4cb9-96ff-ba8780077be3 req-5088f3ff-90ed-44a8-ab97-12a6ed8a5d6b service nova] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.655248] env[62552]: DEBUG nova.compute.manager [req-301827d5-93cc-4cb9-96ff-ba8780077be3 req-5088f3ff-90ed-44a8-ab97-12a6ed8a5d6b service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] No waiting events found dispatching network-vif-unplugged-61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 949.655419] env[62552]: WARNING nova.compute.manager [req-301827d5-93cc-4cb9-96ff-ba8780077be3 req-5088f3ff-90ed-44a8-ab97-12a6ed8a5d6b service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Received unexpected event network-vif-unplugged-61514a33-d741-4aae-876d-b0ecbe6ed6e6 for instance with vm_state shelved and task_state shelving_offloading. [ 949.764805] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.765753] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149111eb-4961-4eb6-a4ba-12d225773081 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.773188] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 949.773463] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff33e587-985b-4560-951a-b5c52f455683 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.836716] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 949.836956] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 949.837160] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleting the datastore file [datastore1] e8e8dec1-4210-44d2-a3f8-76055f2c3b57 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.837445] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a29e14e3-e5c0-49d2-a336-874c785612d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.844325] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 949.844325] env[62552]: value = "task-1239788" [ 949.844325] env[62552]: _type = "Task" [ 949.844325] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.854212] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239788, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.874232] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239785, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.933978] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.934594] env[62552]: DEBUG nova.compute.manager [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 949.938033] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.306s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.939939] env[62552]: INFO nova.compute.claims [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.997763] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239786, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.017799] env[62552]: DEBUG nova.compute.manager [req-734c5233-60aa-43d5-8ada-e26fe3cddc8c req-181bf3a2-cf94-4f80-b64b-17d494fcef81 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Received event network-vif-deleted-9136fa4c-29b1-49d3-ba65-6530ce9441b8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 950.018015] env[62552]: INFO nova.compute.manager [req-734c5233-60aa-43d5-8ada-e26fe3cddc8c req-181bf3a2-cf94-4f80-b64b-17d494fcef81 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Neutron deleted interface 9136fa4c-29b1-49d3-ba65-6530ce9441b8; detaching it from the instance and deleting it from the info cache [ 950.018415] env[62552]: DEBUG nova.network.neutron [req-734c5233-60aa-43d5-8ada-e26fe3cddc8c req-181bf3a2-cf94-4f80-b64b-17d494fcef81 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.357244] env[62552]: DEBUG oslo_vmware.api [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239788, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163884} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.358396] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.358784] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.359141] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.378382] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239785, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.394265] env[62552]: INFO nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted allocations for instance e8e8dec1-4210-44d2-a3f8-76055f2c3b57 [ 950.445150] env[62552]: DEBUG nova.compute.utils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 950.447429] env[62552]: DEBUG nova.compute.manager [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 950.451760] env[62552]: DEBUG nova.network.neutron [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 950.487344] env[62552]: DEBUG nova.network.neutron [-] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.497817] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239786, 'name': ReconfigVM_Task, 'duration_secs': 0.644647} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.498137] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 04186fde-a9ed-415a-bd40-312f0347fcc7/04186fde-a9ed-415a-bd40-312f0347fcc7.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.498789] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c93e556-b66b-4a09-81f5-41974c9f9a90 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.506772] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 950.506772] env[62552]: value = "task-1239789" [ 950.506772] env[62552]: _type = "Task" [ 950.506772] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.516217] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239789, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.523293] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb96d2a8-6abc-402a-8d47-64bce899d63f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.531912] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0a31e0-5f3b-498b-af60-8ba730b72bc9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.544041] env[62552]: DEBUG nova.policy [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80977f84033b4e45a2294b9603f9f358', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f2003df754340ab8e4cf9ca5a21b9a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 950.567676] env[62552]: DEBUG nova.compute.manager [req-734c5233-60aa-43d5-8ada-e26fe3cddc8c req-181bf3a2-cf94-4f80-b64b-17d494fcef81 service nova] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Detach interface failed, port_id=9136fa4c-29b1-49d3-ba65-6530ce9441b8, reason: Instance fbbc360d-9cfd-48f1-80b2-26da2c72c002 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 950.590172] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "80478878-ff82-4ed6-a851-8eb2bec01e22" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.590446] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.590663] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "80478878-ff82-4ed6-a851-8eb2bec01e22-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.590848] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.591033] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.593298] env[62552]: INFO nova.compute.manager [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Terminating instance [ 950.824284] env[62552]: DEBUG nova.network.neutron [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Successfully created port: 19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.880541] env[62552]: DEBUG oslo_vmware.api [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239785, 'name': RemoveSnapshot_Task, 'duration_secs': 1.065575} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.880922] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 950.896588] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.948367] env[62552]: DEBUG nova.compute.manager [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 950.992399] env[62552]: INFO nova.compute.manager [-] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Took 1.49 seconds to deallocate network for instance. [ 951.022036] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239789, 'name': Rename_Task, 'duration_secs': 0.143439} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.025294] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.028884] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97c6c11d-910b-41ed-b251-7d9d4ac1540a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.034999] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 951.034999] env[62552]: value = "task-1239790" [ 951.034999] env[62552]: _type = "Task" [ 951.034999] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.048274] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239790, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.098674] env[62552]: DEBUG nova.compute.manager [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 951.099576] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 951.100447] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27078802-8652-4ec2-bfd8-1a4c65bfe1d7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.107655] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 951.110612] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1218eca9-28bc-4ae5-b72f-4bc7bda7605b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.117430] env[62552]: DEBUG oslo_vmware.api [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 951.117430] env[62552]: value = "task-1239791" [ 951.117430] env[62552]: _type = "Task" [ 951.117430] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.129921] env[62552]: DEBUG oslo_vmware.api [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239791, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.217351] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535e0591-b877-474b-82d1-8658f08a1598 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.225182] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e88ddd-05ed-42c8-9a68-e243f15a266f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.254854] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b16ff96-fa79-4426-bbbd-9b3ddda5e070 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.262350] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef302e09-e66d-4c1b-8ab6-95173355d733 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.276104] env[62552]: DEBUG nova.compute.provider_tree [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.394346] env[62552]: WARNING nova.compute.manager [None req-483e070d-7a4c-4fa1-bd03-d384a6e4a110 tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Image not found during snapshot: nova.exception.ImageNotFound: Image d2991aba-c054-4e7b-90f5-6d2811506639 could not be found. [ 951.504779] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.547710] env[62552]: DEBUG oslo_vmware.api [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239790, 'name': PowerOnVM_Task, 'duration_secs': 0.491186} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.547994] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.548222] env[62552]: INFO nova.compute.manager [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Took 7.42 seconds to spawn the instance on the hypervisor. [ 951.548413] env[62552]: DEBUG nova.compute.manager [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 951.549204] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057dae50-39f9-472c-baf3-8205f44a1b03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.628565] env[62552]: DEBUG oslo_vmware.api [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239791, 'name': PowerOffVM_Task, 'duration_secs': 0.217955} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.628861] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.629048] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 951.629308] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63e2a805-4221-4989-b27a-ee827848fd06 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.715617] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 951.715838] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 951.716086] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Deleting the datastore file [datastore2] 80478878-ff82-4ed6-a851-8eb2bec01e22 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.716655] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b69d4b67-692f-4719-b77b-a8ca34b7ee61 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.723613] env[62552]: DEBUG oslo_vmware.api [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for the task: (returnval){ [ 951.723613] env[62552]: value = "task-1239793" [ 951.723613] env[62552]: _type = "Task" [ 951.723613] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.732516] env[62552]: DEBUG oslo_vmware.api [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.779275] env[62552]: DEBUG nova.scheduler.client.report [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 951.810510] env[62552]: DEBUG nova.compute.manager [req-6782f774-5e69-4d65-b289-c03cba31a097 req-8ea50f8e-4b50-4298-a984-bf25c089a871 service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Received event network-changed-61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 951.810693] env[62552]: DEBUG nova.compute.manager [req-6782f774-5e69-4d65-b289-c03cba31a097 req-8ea50f8e-4b50-4298-a984-bf25c089a871 service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Refreshing instance network info cache due to event network-changed-61514a33-d741-4aae-876d-b0ecbe6ed6e6. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 951.810846] env[62552]: DEBUG oslo_concurrency.lockutils [req-6782f774-5e69-4d65-b289-c03cba31a097 req-8ea50f8e-4b50-4298-a984-bf25c089a871 service nova] Acquiring lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.811024] env[62552]: DEBUG oslo_concurrency.lockutils [req-6782f774-5e69-4d65-b289-c03cba31a097 req-8ea50f8e-4b50-4298-a984-bf25c089a871 service nova] Acquired lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.811754] env[62552]: DEBUG nova.network.neutron [req-6782f774-5e69-4d65-b289-c03cba31a097 req-8ea50f8e-4b50-4298-a984-bf25c089a871 service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Refreshing network info cache for port 61514a33-d741-4aae-876d-b0ecbe6ed6e6 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.959313] env[62552]: DEBUG nova.compute.manager [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 951.975232] env[62552]: DEBUG oslo_concurrency.lockutils [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.988533] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.988828] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.990148] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.990148] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.990148] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.990148] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.990148] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.990148] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.990389] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.990389] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.993545] env[62552]: DEBUG nova.virt.hardware [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.993545] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5afc74af-d670-4f7a-b751-0d98a8c9309e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.001488] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf49d14a-1a2b-477e-8f22-c64cfdf5104a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.069674] env[62552]: INFO nova.compute.manager [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Took 19.16 seconds to build instance. [ 952.221559] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "c5ed17ee-7421-49d7-975f-1515de1e4b01" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.221559] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "c5ed17ee-7421-49d7-975f-1515de1e4b01" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.221559] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "c5ed17ee-7421-49d7-975f-1515de1e4b01-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.221559] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "c5ed17ee-7421-49d7-975f-1515de1e4b01-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.221559] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "c5ed17ee-7421-49d7-975f-1515de1e4b01-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.223971] env[62552]: INFO nova.compute.manager [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Terminating instance [ 952.238758] env[62552]: DEBUG oslo_vmware.api [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Task: {'id': task-1239793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149521} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.239562] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.240107] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 952.240455] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 952.240864] env[62552]: INFO nova.compute.manager [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Took 1.14 seconds to destroy the instance on the hypervisor. [ 952.241090] env[62552]: DEBUG oslo.service.loopingcall [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.241305] env[62552]: DEBUG nova.compute.manager [-] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 952.241410] env[62552]: DEBUG nova.network.neutron [-] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 952.285120] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.285659] env[62552]: DEBUG nova.compute.manager [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 952.288490] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.392s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.288711] env[62552]: DEBUG nova.objects.instance [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'resources' on Instance uuid e8e8dec1-4210-44d2-a3f8-76055f2c3b57 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.321095] env[62552]: DEBUG nova.compute.manager [req-d2858996-78f0-4f14-92b3-f50e9812063b req-0a048fb5-0b53-4e9a-a9f3-95e2f8e7fefe service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Received event network-vif-plugged-19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 952.321364] env[62552]: DEBUG oslo_concurrency.lockutils [req-d2858996-78f0-4f14-92b3-f50e9812063b req-0a048fb5-0b53-4e9a-a9f3-95e2f8e7fefe service nova] Acquiring lock "79166f2c-f864-4d8a-b1dc-e176710400cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.321840] env[62552]: DEBUG oslo_concurrency.lockutils [req-d2858996-78f0-4f14-92b3-f50e9812063b req-0a048fb5-0b53-4e9a-a9f3-95e2f8e7fefe service nova] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.321840] env[62552]: DEBUG oslo_concurrency.lockutils [req-d2858996-78f0-4f14-92b3-f50e9812063b req-0a048fb5-0b53-4e9a-a9f3-95e2f8e7fefe service nova] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.322660] env[62552]: DEBUG nova.compute.manager [req-d2858996-78f0-4f14-92b3-f50e9812063b req-0a048fb5-0b53-4e9a-a9f3-95e2f8e7fefe service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] No waiting events found dispatching network-vif-plugged-19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 952.322891] env[62552]: WARNING nova.compute.manager [req-d2858996-78f0-4f14-92b3-f50e9812063b req-0a048fb5-0b53-4e9a-a9f3-95e2f8e7fefe service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Received unexpected event network-vif-plugged-19a87259-1f7a-428c-8790-b74e27fde6b6 for instance with vm_state building and task_state spawning. [ 952.458590] env[62552]: DEBUG nova.network.neutron [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Successfully updated port: 19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 952.572294] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9537d317-1a06-4dbc-a3b0-1a40d6a7d4b7 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.684s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.732559] env[62552]: DEBUG nova.compute.manager [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 952.732804] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.733761] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04501f50-7353-4427-b014-332926b3a0c9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.742359] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.742621] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c04e691-e1bf-4162-9030-d2699b6fffac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.748487] env[62552]: DEBUG oslo_vmware.api [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 952.748487] env[62552]: value = "task-1239794" [ 952.748487] env[62552]: _type = "Task" [ 952.748487] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.756907] env[62552]: DEBUG oslo_vmware.api [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239794, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.793684] env[62552]: DEBUG nova.compute.utils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 952.795332] env[62552]: DEBUG nova.objects.instance [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'numa_topology' on Instance uuid e8e8dec1-4210-44d2-a3f8-76055f2c3b57 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.796496] env[62552]: DEBUG nova.compute.manager [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 952.796664] env[62552]: DEBUG nova.network.neutron [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 952.839689] env[62552]: DEBUG nova.policy [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6515846d33e44bca942767f9e1d650b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2f57f72de0c4adc9754458f4e7040b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 952.965725] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.965916] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.966092] env[62552]: DEBUG nova.network.neutron [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.167891] env[62552]: DEBUG nova.network.neutron [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Successfully created port: bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 953.259404] env[62552]: DEBUG oslo_vmware.api [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239794, 'name': PowerOffVM_Task, 'duration_secs': 0.244528} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.262899] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.263096] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.263368] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-503de91a-fe97-42c3-9fc2-b6265635a10f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.297414] env[62552]: DEBUG nova.compute.manager [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 953.301357] env[62552]: DEBUG nova.objects.base [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 953.325453] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.325583] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.325783] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleting the datastore file [datastore2] c5ed17ee-7421-49d7-975f-1515de1e4b01 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.325927] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c3b7d7a-4d7e-4438-86a8-441e938ba22a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.332978] env[62552]: DEBUG oslo_vmware.api [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for the task: (returnval){ [ 953.332978] env[62552]: value = "task-1239796" [ 953.332978] env[62552]: _type = "Task" [ 953.332978] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.347129] env[62552]: DEBUG oslo_vmware.api [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239796, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.405234] env[62552]: DEBUG nova.network.neutron [req-6782f774-5e69-4d65-b289-c03cba31a097 req-8ea50f8e-4b50-4298-a984-bf25c089a871 service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Updated VIF entry in instance network info cache for port 61514a33-d741-4aae-876d-b0ecbe6ed6e6. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.405851] env[62552]: DEBUG nova.network.neutron [req-6782f774-5e69-4d65-b289-c03cba31a097 req-8ea50f8e-4b50-4298-a984-bf25c089a871 service nova] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Updating instance_info_cache with network_info: [{"id": "61514a33-d741-4aae-876d-b0ecbe6ed6e6", "address": "fa:16:3e:b9:cf:bd", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": null, "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap61514a33-d7", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.521214] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d059e8c1-b4cf-4a1b-84a0-6ed2099ba3ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.535237] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e824d7fa-3b87-4b69-afab-ee1ea0555151 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.572110] env[62552]: DEBUG nova.network.neutron [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 953.574921] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d93aeb1-b91b-47c6-95b8-3228d34c1e21 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.584090] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36d8148-3b3f-436f-8888-56fd3227c6da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.603572] env[62552]: DEBUG nova.compute.provider_tree [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 953.616631] env[62552]: DEBUG nova.compute.manager [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Stashing vm_state: active {{(pid=62552) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 953.846375] env[62552]: DEBUG oslo_vmware.api [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Task: {'id': task-1239796, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144445} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.846814] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.847075] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.847292] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.847545] env[62552]: INFO nova.compute.manager [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Took 1.11 seconds to destroy the instance on the hypervisor. [ 953.847938] env[62552]: DEBUG oslo.service.loopingcall [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.848383] env[62552]: DEBUG nova.compute.manager [-] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 953.848505] env[62552]: DEBUG nova.network.neutron [-] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.910182] env[62552]: DEBUG oslo_concurrency.lockutils [req-6782f774-5e69-4d65-b289-c03cba31a097 req-8ea50f8e-4b50-4298-a984-bf25c089a871 service nova] Releasing lock "refresh_cache-e8e8dec1-4210-44d2-a3f8-76055f2c3b57" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.014245] env[62552]: DEBUG nova.compute.manager [req-370c3910-89a2-44dd-a6e5-82fb18f98417 req-fb45364d-49f9-4657-8e9a-e3e2bb1e4641 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Received event network-vif-deleted-76bce90f-8d24-4e1e-8562-f2790b183627 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 954.014245] env[62552]: INFO nova.compute.manager [req-370c3910-89a2-44dd-a6e5-82fb18f98417 req-fb45364d-49f9-4657-8e9a-e3e2bb1e4641 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Neutron deleted interface 76bce90f-8d24-4e1e-8562-f2790b183627; detaching it from the instance and deleting it from the info cache [ 954.014245] env[62552]: DEBUG nova.network.neutron [req-370c3910-89a2-44dd-a6e5-82fb18f98417 req-fb45364d-49f9-4657-8e9a-e3e2bb1e4641 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.060674] env[62552]: DEBUG nova.network.neutron [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating instance_info_cache with network_info: [{"id": "19a87259-1f7a-428c-8790-b74e27fde6b6", "address": "fa:16:3e:04:c8:af", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19a87259-1f", "ovs_interfaceid": "19a87259-1f7a-428c-8790-b74e27fde6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.142314] env[62552]: ERROR nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [req-b5834865-53fa-44b8-853f-0683200d19d2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b5834865-53fa-44b8-853f-0683200d19d2"}]} [ 954.150018] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.173906] env[62552]: DEBUG nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 954.192388] env[62552]: DEBUG nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 954.194747] env[62552]: DEBUG nova.compute.provider_tree [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 954.208267] env[62552]: DEBUG nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 954.238621] env[62552]: DEBUG nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 954.309815] env[62552]: DEBUG nova.compute.manager [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 954.335284] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.335586] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.335760] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.335947] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.336368] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.336605] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.336867] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.337213] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.337277] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.337503] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.337752] env[62552]: DEBUG nova.virt.hardware [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.338811] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea2e36d-9415-48e1-aaa2-72e003d56c6c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.351708] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f31193a-4588-4bd6-882e-9066b55acb2d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.412549] env[62552]: DEBUG nova.network.neutron [-] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.477266] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7a8ec1-ab99-45f9-9016-b24204fb4594 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.485792] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32d1c6b-3408-4464-abe2-9539462b5735 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.518707] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-109df6a2-8dd9-4316-a94d-f93294cd02aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.521357] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb9a5ab-9d76-45a1-935f-bceba3c91b8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.530895] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60098b9-379f-401c-baae-832058826e45 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.537401] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a2f51e-aa89-4429-9fff-8005a105a977 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.560934] env[62552]: DEBUG nova.compute.provider_tree [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 954.570386] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.570707] env[62552]: DEBUG nova.compute.manager [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Instance network_info: |[{"id": "19a87259-1f7a-428c-8790-b74e27fde6b6", "address": "fa:16:3e:04:c8:af", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19a87259-1f", "ovs_interfaceid": "19a87259-1f7a-428c-8790-b74e27fde6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 954.571092] env[62552]: DEBUG nova.compute.manager [req-370c3910-89a2-44dd-a6e5-82fb18f98417 req-fb45364d-49f9-4657-8e9a-e3e2bb1e4641 service nova] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Detach interface failed, port_id=76bce90f-8d24-4e1e-8562-f2790b183627, reason: Instance 80478878-ff82-4ed6-a851-8eb2bec01e22 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 954.572378] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:c8:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '838c9497-35dd-415e-96c7-8dc21b0cd4b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19a87259-1f7a-428c-8790-b74e27fde6b6', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.579905] env[62552]: DEBUG oslo.service.loopingcall [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.580405] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 954.580635] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3f3fb09-5bfa-4a28-b973-db8de2382242 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.602907] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.602907] env[62552]: value = "task-1239797" [ 954.602907] env[62552]: _type = "Task" [ 954.602907] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.614381] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239797, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.703110] env[62552]: DEBUG nova.compute.manager [req-6ac76cd1-afbb-42b2-90e6-7a318901bac6 req-dd921594-c52b-466a-9a3f-4f8f4515c0ab service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Received event network-changed-19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 954.703110] env[62552]: DEBUG nova.compute.manager [req-6ac76cd1-afbb-42b2-90e6-7a318901bac6 req-dd921594-c52b-466a-9a3f-4f8f4515c0ab service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Refreshing instance network info cache due to event network-changed-19a87259-1f7a-428c-8790-b74e27fde6b6. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 954.703728] env[62552]: DEBUG oslo_concurrency.lockutils [req-6ac76cd1-afbb-42b2-90e6-7a318901bac6 req-dd921594-c52b-466a-9a3f-4f8f4515c0ab service nova] Acquiring lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.703728] env[62552]: DEBUG oslo_concurrency.lockutils [req-6ac76cd1-afbb-42b2-90e6-7a318901bac6 req-dd921594-c52b-466a-9a3f-4f8f4515c0ab service nova] Acquired lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.703728] env[62552]: DEBUG nova.network.neutron [req-6ac76cd1-afbb-42b2-90e6-7a318901bac6 req-dd921594-c52b-466a-9a3f-4f8f4515c0ab service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Refreshing network info cache for port 19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 954.711326] env[62552]: DEBUG nova.network.neutron [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Successfully updated port: bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 954.907175] env[62552]: DEBUG nova.network.neutron [-] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.917331] env[62552]: INFO nova.compute.manager [-] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Took 2.68 seconds to deallocate network for instance. [ 955.091612] env[62552]: ERROR nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [req-08766fac-37a7-45ef-abb3-6d0d62ec8967] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-08766fac-37a7-45ef-abb3-6d0d62ec8967"}]} [ 955.107959] env[62552]: DEBUG nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 955.116526] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239797, 'name': CreateVM_Task, 'duration_secs': 0.398938} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.116674] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 955.117378] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.117551] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.117879] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.118160] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67d3bc58-013e-4144-8a72-8209dfa71dec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.124744] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 955.124744] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5288d6f1-8a0d-cef0-40e4-b1e7053113c4" [ 955.124744] env[62552]: _type = "Task" [ 955.124744] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.126413] env[62552]: DEBUG nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 955.126533] env[62552]: DEBUG nova.compute.provider_tree [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 955.137012] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5288d6f1-8a0d-cef0-40e4-b1e7053113c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.138024] env[62552]: DEBUG nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 955.157515] env[62552]: DEBUG nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 955.217654] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.217654] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.217654] env[62552]: DEBUG nova.network.neutron [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.306020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.306175] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.309796] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquiring lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.310081] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.384565] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd835feb-116a-47a8-b722-020ecd5afdaf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.392511] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29897217-858a-4e76-ab41-1e10339615e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.424029] env[62552]: INFO nova.compute.manager [-] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Took 1.58 seconds to deallocate network for instance. [ 955.424934] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.430776] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31c1056-808a-4022-a726-a04ba21f9373 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.433989] env[62552]: DEBUG nova.network.neutron [req-6ac76cd1-afbb-42b2-90e6-7a318901bac6 req-dd921594-c52b-466a-9a3f-4f8f4515c0ab service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updated VIF entry in instance network info cache for port 19a87259-1f7a-428c-8790-b74e27fde6b6. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 955.434327] env[62552]: DEBUG nova.network.neutron [req-6ac76cd1-afbb-42b2-90e6-7a318901bac6 req-dd921594-c52b-466a-9a3f-4f8f4515c0ab service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating instance_info_cache with network_info: [{"id": "19a87259-1f7a-428c-8790-b74e27fde6b6", "address": "fa:16:3e:04:c8:af", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19a87259-1f", "ovs_interfaceid": "19a87259-1f7a-428c-8790-b74e27fde6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.442192] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180319f7-3e03-4091-a310-249bd10816d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.456817] env[62552]: DEBUG nova.compute.provider_tree [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 955.639573] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5288d6f1-8a0d-cef0-40e4-b1e7053113c4, 'name': SearchDatastore_Task, 'duration_secs': 0.040619} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.639891] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.640149] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.640387] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.640537] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.640796] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.640976] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fed8ec4b-1877-4819-a686-cb6bd3e86a34 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.653916] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.654112] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 955.654802] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61f9ae9e-c7f1-4fb3-8e17-b67a73e94343 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.659725] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 955.659725] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52af519c-72a4-1e67-2032-648d8782738a" [ 955.659725] env[62552]: _type = "Task" [ 955.659725] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.666745] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52af519c-72a4-1e67-2032-648d8782738a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.748827] env[62552]: DEBUG nova.network.neutron [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 955.809558] env[62552]: DEBUG nova.compute.manager [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 955.818034] env[62552]: DEBUG nova.compute.manager [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 955.887782] env[62552]: DEBUG nova.network.neutron [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.931304] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.936855] env[62552]: DEBUG oslo_concurrency.lockutils [req-6ac76cd1-afbb-42b2-90e6-7a318901bac6 req-dd921594-c52b-466a-9a3f-4f8f4515c0ab service nova] Releasing lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.988785] env[62552]: DEBUG nova.scheduler.client.report [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 955.989238] env[62552]: DEBUG nova.compute.provider_tree [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 107 to 108 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 955.989512] env[62552]: DEBUG nova.compute.provider_tree [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 956.109814] env[62552]: DEBUG nova.compute.manager [req-1ce43303-ec36-4c5e-8b39-b41ed5eeebb1 req-42ae7dd6-b474-4924-b5e6-d2a2a80f16c0 service nova] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Received event network-vif-deleted-8018d47b-da5e-40c6-b217-a8b685b44983 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 956.172511] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52af519c-72a4-1e67-2032-648d8782738a, 'name': SearchDatastore_Task, 'duration_secs': 0.042248} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.172511] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c72d6f1-1e33-4c7d-a69c-8a94792d1bb7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.177016] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 956.177016] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520e57b4-4fc2-d78d-f3c9-b7dd869139a1" [ 956.177016] env[62552]: _type = "Task" [ 956.177016] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.183997] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520e57b4-4fc2-d78d-f3c9-b7dd869139a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.332723] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.334545] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.391503] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.391867] env[62552]: DEBUG nova.compute.manager [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Instance network_info: |[{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 956.392333] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:2a:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0a3f99df-d1bc-4a37-a048-263445d4a7b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb66aa2b-bd30-454a-b71b-da2a0285cef1', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.399681] env[62552]: DEBUG oslo.service.loopingcall [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.399902] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 956.400152] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e68b9f2-74ef-453b-baef-131369d41c4f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.420027] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.420027] env[62552]: value = "task-1239798" [ 956.420027] env[62552]: _type = "Task" [ 956.420027] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.429992] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239798, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.495441] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.207s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.498115] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.994s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.498360] env[62552]: DEBUG nova.objects.instance [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'resources' on Instance uuid fbbc360d-9cfd-48f1-80b2-26da2c72c002 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.687041] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520e57b4-4fc2-d78d-f3c9-b7dd869139a1, 'name': SearchDatastore_Task, 'duration_secs': 0.009692} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.687041] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.687443] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 79166f2c-f864-4d8a-b1dc-e176710400cb/79166f2c-f864-4d8a-b1dc-e176710400cb.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 956.687443] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3d6ec8e-cf2f-4789-8cd0-3668d571dc3c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.694775] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 956.694775] env[62552]: value = "task-1239799" [ 956.694775] env[62552]: _type = "Task" [ 956.694775] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.703203] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239799, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.732253] env[62552]: DEBUG nova.compute.manager [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received event network-vif-plugged-bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 956.732448] env[62552]: DEBUG oslo_concurrency.lockutils [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] Acquiring lock "6a4573f0-7152-4462-bb72-58a45b2cbd97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.732703] env[62552]: DEBUG oslo_concurrency.lockutils [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.732920] env[62552]: DEBUG oslo_concurrency.lockutils [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.733187] env[62552]: DEBUG nova.compute.manager [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] No waiting events found dispatching network-vif-plugged-bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 956.733404] env[62552]: WARNING nova.compute.manager [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received unexpected event network-vif-plugged-bb66aa2b-bd30-454a-b71b-da2a0285cef1 for instance with vm_state building and task_state spawning. [ 956.733637] env[62552]: DEBUG nova.compute.manager [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 956.733844] env[62552]: DEBUG nova.compute.manager [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing instance network info cache due to event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 956.734127] env[62552]: DEBUG oslo_concurrency.lockutils [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] Acquiring lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.734341] env[62552]: DEBUG oslo_concurrency.lockutils [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] Acquired lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.734600] env[62552]: DEBUG nova.network.neutron [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 956.932773] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239798, 'name': CreateVM_Task, 'duration_secs': 0.328114} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.932967] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 956.933788] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.934035] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.934502] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.934760] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-356bca4d-b583-4933-add3-4ff10626742d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.941337] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 956.941337] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e2ff3e-cf9d-4a44-ddce-ef3e031948c5" [ 956.941337] env[62552]: _type = "Task" [ 956.941337] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.950248] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e2ff3e-cf9d-4a44-ddce-ef3e031948c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.007095] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f00d31f4-ad04-4ff7-ac36-f4e0b9997702 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 28.496s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.008612] env[62552]: DEBUG oslo_concurrency.lockutils [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.033s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.008943] env[62552]: DEBUG oslo_concurrency.lockutils [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.009170] env[62552]: DEBUG oslo_concurrency.lockutils [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.009344] env[62552]: DEBUG oslo_concurrency.lockutils [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.011851] env[62552]: INFO nova.compute.manager [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Terminating instance [ 957.209329] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239799, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447744} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.209647] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 79166f2c-f864-4d8a-b1dc-e176710400cb/79166f2c-f864-4d8a-b1dc-e176710400cb.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 957.209873] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 957.210282] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94745451-f30d-4500-ac8e-c955ca4f10b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.217036] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 957.217036] env[62552]: value = "task-1239800" [ 957.217036] env[62552]: _type = "Task" [ 957.217036] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.225275] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239800, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.253720] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd451a36-5bef-4c13-ae05-d2ab5a31321e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.261093] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a37f255-8773-44ae-9ce9-83d21c1e6041 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.294675] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b464c83e-72db-463f-b8ef-c60ee6559dd5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.302877] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e576298-ca1a-48e2-83ff-50b8478602d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.316522] env[62552]: DEBUG nova.compute.provider_tree [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 957.451653] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e2ff3e-cf9d-4a44-ddce-ef3e031948c5, 'name': SearchDatastore_Task, 'duration_secs': 0.050616} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.451974] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.452239] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.452480] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.452634] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.452818] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.453104] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9149552d-cfa8-47cc-8df9-fb372ab816be {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.458766] env[62552]: DEBUG nova.network.neutron [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updated VIF entry in instance network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 957.459974] env[62552]: DEBUG nova.network.neutron [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.461534] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.461718] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.462608] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7eb1b3bd-b6ad-46bf-956e-62f8af7b334b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.468227] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 957.468227] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5214edc3-9bea-cdaa-5b76-31028b807ad7" [ 957.468227] env[62552]: _type = "Task" [ 957.468227] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.476866] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5214edc3-9bea-cdaa-5b76-31028b807ad7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.516840] env[62552]: DEBUG nova.compute.manager [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 957.517022] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 957.517300] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5580190c-67c5-45a9-9a45-29511f68d5e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.530023] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5061f13-5454-4c7d-bfd2-f8a3a7935d7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.558918] env[62552]: WARNING nova.virt.vmwareapi.vmops [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8e8dec1-4210-44d2-a3f8-76055f2c3b57 could not be found. [ 957.558918] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.558918] env[62552]: INFO nova.compute.manager [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Took 0.04 seconds to destroy the instance on the hypervisor. [ 957.558918] env[62552]: DEBUG oslo.service.loopingcall [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.559148] env[62552]: DEBUG nova.compute.manager [-] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 957.559148] env[62552]: DEBUG nova.network.neutron [-] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.729471] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239800, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067113} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.729830] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.730701] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6838b7e5-0034-4516-bdbb-ce462e985308 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.752472] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 79166f2c-f864-4d8a-b1dc-e176710400cb/79166f2c-f864-4d8a-b1dc-e176710400cb.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.753341] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-961d229a-b6ad-4095-adc0-afa81408fd51 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.775117] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 957.775117] env[62552]: value = "task-1239801" [ 957.775117] env[62552]: _type = "Task" [ 957.775117] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.785696] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239801, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.846058] env[62552]: DEBUG nova.scheduler.client.report [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 108 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 957.846378] env[62552]: DEBUG nova.compute.provider_tree [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 108 to 109 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 957.846566] env[62552]: DEBUG nova.compute.provider_tree [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 957.963241] env[62552]: DEBUG oslo_concurrency.lockutils [req-eb7b9016-56d1-4e56-ba30-b68fb3ce8e05 req-42bbe953-c164-4521-93ef-e07c7c78a341 service nova] Releasing lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.977831] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5214edc3-9bea-cdaa-5b76-31028b807ad7, 'name': SearchDatastore_Task, 'duration_secs': 0.009192} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.978617] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f642f646-ab83-4d2b-abf6-e1f069d82596 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.983358] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 957.983358] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ff4163-4e53-cec8-a7c2-afdd7729a695" [ 957.983358] env[62552]: _type = "Task" [ 957.983358] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.990376] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ff4163-4e53-cec8-a7c2-afdd7729a695, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.280822] env[62552]: DEBUG nova.network.neutron [-] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.285536] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239801, 'name': ReconfigVM_Task, 'duration_secs': 0.268841} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.286053] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 79166f2c-f864-4d8a-b1dc-e176710400cb/79166f2c-f864-4d8a-b1dc-e176710400cb.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.286683] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-973d2fe6-ad31-4bf5-b840-325ed4497a9d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.293887] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 958.293887] env[62552]: value = "task-1239802" [ 958.293887] env[62552]: _type = "Task" [ 958.293887] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.302284] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239802, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.355109] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.854s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.355109] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.208s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.377103] env[62552]: INFO nova.scheduler.client.report [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleted allocations for instance fbbc360d-9cfd-48f1-80b2-26da2c72c002 [ 958.494250] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ff4163-4e53-cec8-a7c2-afdd7729a695, 'name': SearchDatastore_Task, 'duration_secs': 0.016093} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.494546] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.494819] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6a4573f0-7152-4462-bb72-58a45b2cbd97/6a4573f0-7152-4462-bb72-58a45b2cbd97.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 958.495107] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f0d2c38-8d74-42b5-a9fc-11b6fc763ad0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.500888] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 958.500888] env[62552]: value = "task-1239803" [ 958.500888] env[62552]: _type = "Task" [ 958.500888] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.508620] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239803, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.786887] env[62552]: INFO nova.compute.manager [-] [instance: e8e8dec1-4210-44d2-a3f8-76055f2c3b57] Took 1.23 seconds to deallocate network for instance. [ 958.806365] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239802, 'name': Rename_Task, 'duration_secs': 0.167714} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.806684] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.806962] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4789130a-7b44-4a4f-8ded-a59eb49061b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.813690] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 958.813690] env[62552]: value = "task-1239804" [ 958.813690] env[62552]: _type = "Task" [ 958.813690] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.822255] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239804, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.860847] env[62552]: INFO nova.compute.claims [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.885064] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3859bed-da58-4c35-8d5c-1e953677f6bd tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "fbbc360d-9cfd-48f1-80b2-26da2c72c002" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.060s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.010968] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239803, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470288} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.011333] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6a4573f0-7152-4462-bb72-58a45b2cbd97/6a4573f0-7152-4462-bb72-58a45b2cbd97.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 959.011636] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 959.011926] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df22742d-6b8f-45d5-bdd7-6f6a90c62231 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.017674] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 959.017674] env[62552]: value = "task-1239805" [ 959.017674] env[62552]: _type = "Task" [ 959.017674] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.025285] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239805, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.323719] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239804, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.368051] env[62552]: INFO nova.compute.resource_tracker [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating resource usage from migration 49939264-6417-4e73-9ac7-4f05edf037aa [ 959.529915] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239805, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082999} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.530277] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.531201] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a1e94d-1fe2-45b5-9dc9-033dd7e0dd3e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.556178] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 6a4573f0-7152-4462-bb72-58a45b2cbd97/6a4573f0-7152-4462-bb72-58a45b2cbd97.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.559434] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4b7c7ea-ae66-4d9d-9598-14cc36517ad5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.579888] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 959.579888] env[62552]: value = "task-1239806" [ 959.579888] env[62552]: _type = "Task" [ 959.579888] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.591332] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239806, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.598894] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c440bfc-783b-4571-b8ea-7931f998b0e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.606605] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3f5301-1962-448d-bd3d-f241cf1a2835 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.637886] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2779df-bdd2-4d32-83b7-536dc92f7a65 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.645914] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58bb7ec3-d021-486e-8da9-7c038e27f060 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.660174] env[62552]: DEBUG nova.compute.provider_tree [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.818793] env[62552]: DEBUG oslo_concurrency.lockutils [None req-58f12b11-40f9-4d43-b088-3201fa97c714 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e8e8dec1-4210-44d2-a3f8-76055f2c3b57" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.810s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.825345] env[62552]: DEBUG oslo_vmware.api [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239804, 'name': PowerOnVM_Task, 'duration_secs': 0.53222} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.825642] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.826328] env[62552]: INFO nova.compute.manager [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Took 7.87 seconds to spawn the instance on the hypervisor. [ 959.826328] env[62552]: DEBUG nova.compute.manager [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 959.826919] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fce662-98d5-4810-b767-191666966d1e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.091824] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239806, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.163471] env[62552]: DEBUG nova.scheduler.client.report [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 960.344234] env[62552]: INFO nova.compute.manager [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Took 19.28 seconds to build instance. [ 960.526023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "0418260a-aa27-4955-ab15-b180ec04f0b3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.526328] env[62552]: DEBUG oslo_concurrency.lockutils [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.591073] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239806, 'name': ReconfigVM_Task, 'duration_secs': 0.68904} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.591628] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 6a4573f0-7152-4462-bb72-58a45b2cbd97/6a4573f0-7152-4462-bb72-58a45b2cbd97.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 960.592070] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8bf42e12-5c75-4726-b052-49e008990f4c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.599509] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 960.599509] env[62552]: value = "task-1239807" [ 960.599509] env[62552]: _type = "Task" [ 960.599509] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.612031] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239807, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.626027] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.626365] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.668326] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.313s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.668521] env[62552]: INFO nova.compute.manager [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Migrating [ 960.678238] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.253s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.678473] env[62552]: DEBUG nova.objects.instance [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lazy-loading 'resources' on Instance uuid 80478878-ff82-4ed6-a851-8eb2bec01e22 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.846709] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12a54401-8e4c-467f-a6a9-c1969b175d87 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.791s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.029754] env[62552]: INFO nova.compute.manager [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Detaching volume 2d394306-c3fc-458b-bd15-0d1b8649225e [ 961.073557] env[62552]: INFO nova.virt.block_device [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Attempting to driver detach volume 2d394306-c3fc-458b-bd15-0d1b8649225e from mountpoint /dev/sdb [ 961.074209] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 961.074456] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267414', 'volume_id': '2d394306-c3fc-458b-bd15-0d1b8649225e', 'name': 'volume-2d394306-c3fc-458b-bd15-0d1b8649225e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0418260a-aa27-4955-ab15-b180ec04f0b3', 'attached_at': '', 'detached_at': '', 'volume_id': '2d394306-c3fc-458b-bd15-0d1b8649225e', 'serial': '2d394306-c3fc-458b-bd15-0d1b8649225e'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 961.075379] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228ab3c7-058f-4edd-ae80-4e6a71ed79d7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.098793] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8de6373-9154-406e-91c7-8bfa3232eb7a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.113470] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63650a94-1959-4488-b3c3-4e81f33dd985 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.116069] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239807, 'name': Rename_Task, 'duration_secs': 0.349541} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.116432] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.117159] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85ccc216-66ba-4dc4-a68e-58017124e3ea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.135708] env[62552]: DEBUG nova.compute.manager [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 961.139209] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dde5ba9-4710-410a-9517-3f4bac2d9632 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.143657] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 961.143657] env[62552]: value = "task-1239808" [ 961.143657] env[62552]: _type = "Task" [ 961.143657] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.162246] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] The volume has not been displaced from its original location: [datastore2] volume-2d394306-c3fc-458b-bd15-0d1b8649225e/volume-2d394306-c3fc-458b-bd15-0d1b8649225e.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 961.167840] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Reconfiguring VM instance instance-00000031 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 961.168937] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-308b2327-a3b6-4fc5-99a5-8be6d300f75a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.188850] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239808, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.189315] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.189549] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.189758] env[62552]: DEBUG nova.network.neutron [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 961.196891] env[62552]: DEBUG oslo_vmware.api [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 961.196891] env[62552]: value = "task-1239809" [ 961.196891] env[62552]: _type = "Task" [ 961.196891] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.207230] env[62552]: DEBUG oslo_vmware.api [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239809, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.422015] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58414b7b-4f97-4bf8-98bb-e70d2d715128 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.430137] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d06f47-6877-41cb-b731-683f889534e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.463259] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01417f08-2388-42e1-9635-35882fafd4e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.469822] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c0b797-f19d-4e7b-bf0f-457904229a18 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.484259] env[62552]: DEBUG nova.compute.provider_tree [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.517609] env[62552]: DEBUG nova.compute.manager [req-62690f86-1898-4939-891e-32021cb53e0e req-8d34fc12-766e-4817-8e0d-a7a81e6859d9 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Received event network-changed-19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 961.517825] env[62552]: DEBUG nova.compute.manager [req-62690f86-1898-4939-891e-32021cb53e0e req-8d34fc12-766e-4817-8e0d-a7a81e6859d9 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Refreshing instance network info cache due to event network-changed-19a87259-1f7a-428c-8790-b74e27fde6b6. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 961.518026] env[62552]: DEBUG oslo_concurrency.lockutils [req-62690f86-1898-4939-891e-32021cb53e0e req-8d34fc12-766e-4817-8e0d-a7a81e6859d9 service nova] Acquiring lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.518177] env[62552]: DEBUG oslo_concurrency.lockutils [req-62690f86-1898-4939-891e-32021cb53e0e req-8d34fc12-766e-4817-8e0d-a7a81e6859d9 service nova] Acquired lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.518340] env[62552]: DEBUG nova.network.neutron [req-62690f86-1898-4939-891e-32021cb53e0e req-8d34fc12-766e-4817-8e0d-a7a81e6859d9 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Refreshing network info cache for port 19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 961.658625] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239808, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.662298] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.707887] env[62552]: DEBUG oslo_vmware.api [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239809, 'name': ReconfigVM_Task, 'duration_secs': 0.308223} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.707887] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Reconfigured VM instance instance-00000031 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 961.711909] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0845e0b-152b-4137-81fb-e1521302b932 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.729807] env[62552]: DEBUG oslo_vmware.api [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 961.729807] env[62552]: value = "task-1239810" [ 961.729807] env[62552]: _type = "Task" [ 961.729807] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.739303] env[62552]: DEBUG oslo_vmware.api [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239810, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.917449] env[62552]: DEBUG nova.network.neutron [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance_info_cache with network_info: [{"id": "04e39cd5-39ae-4671-b715-e300dec6021d", "address": "fa:16:3e:6e:65:88", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04e39cd5-39", "ovs_interfaceid": "04e39cd5-39ae-4671-b715-e300dec6021d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.940044] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "505d7f8e-c27f-487e-98d8-c840a526d40e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.940388] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "505d7f8e-c27f-487e-98d8-c840a526d40e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.987230] env[62552]: DEBUG nova.scheduler.client.report [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 962.159305] env[62552]: DEBUG oslo_vmware.api [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239808, 'name': PowerOnVM_Task, 'duration_secs': 0.813002} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.159664] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.159885] env[62552]: INFO nova.compute.manager [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Took 7.85 seconds to spawn the instance on the hypervisor. [ 962.160093] env[62552]: DEBUG nova.compute.manager [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 962.160888] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404bc6ab-ca06-4055-9e2f-225ab2417818 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.239060] env[62552]: DEBUG oslo_vmware.api [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239810, 'name': ReconfigVM_Task, 'duration_secs': 0.141168} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.239385] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267414', 'volume_id': '2d394306-c3fc-458b-bd15-0d1b8649225e', 'name': 'volume-2d394306-c3fc-458b-bd15-0d1b8649225e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0418260a-aa27-4955-ab15-b180ec04f0b3', 'attached_at': '', 'detached_at': '', 'volume_id': '2d394306-c3fc-458b-bd15-0d1b8649225e', 'serial': '2d394306-c3fc-458b-bd15-0d1b8649225e'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 962.245548] env[62552]: DEBUG nova.network.neutron [req-62690f86-1898-4939-891e-32021cb53e0e req-8d34fc12-766e-4817-8e0d-a7a81e6859d9 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updated VIF entry in instance network info cache for port 19a87259-1f7a-428c-8790-b74e27fde6b6. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 962.245889] env[62552]: DEBUG nova.network.neutron [req-62690f86-1898-4939-891e-32021cb53e0e req-8d34fc12-766e-4817-8e0d-a7a81e6859d9 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating instance_info_cache with network_info: [{"id": "19a87259-1f7a-428c-8790-b74e27fde6b6", "address": "fa:16:3e:04:c8:af", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19a87259-1f", "ovs_interfaceid": "19a87259-1f7a-428c-8790-b74e27fde6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.420759] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.443027] env[62552]: DEBUG nova.compute.manager [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 962.494291] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.496871] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.565s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.497139] env[62552]: DEBUG nova.objects.instance [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lazy-loading 'resources' on Instance uuid c5ed17ee-7421-49d7-975f-1515de1e4b01 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.517589] env[62552]: INFO nova.scheduler.client.report [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Deleted allocations for instance 80478878-ff82-4ed6-a851-8eb2bec01e22 [ 962.677319] env[62552]: INFO nova.compute.manager [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Took 14.08 seconds to build instance. [ 962.748548] env[62552]: DEBUG oslo_concurrency.lockutils [req-62690f86-1898-4939-891e-32021cb53e0e req-8d34fc12-766e-4817-8e0d-a7a81e6859d9 service nova] Releasing lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.784494] env[62552]: DEBUG nova.objects.instance [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'flavor' on Instance uuid 0418260a-aa27-4955-ab15-b180ec04f0b3 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.962787] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.026517] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6576d0af-4993-4d5a-bf39-d5883c1dc8aa tempest-VolumesAdminNegativeTest-1585057116 tempest-VolumesAdminNegativeTest-1585057116-project-member] Lock "80478878-ff82-4ed6-a851-8eb2bec01e22" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.436s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.179766] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0efe86e-5171-438f-85eb-834e8354d45f tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.587s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.226545] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c93cabf-35d2-4332-be9a-7508507dc85b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.234200] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35a394e-e0cb-4b1f-86b7-36ebb9ad4c26 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.266152] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2f714a-09e0-41c2-900d-b99cef6ba290 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.273577] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb69b3e8-2ad6-4fda-834f-1723097a8ba1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.288680] env[62552]: DEBUG nova.compute.provider_tree [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.726116] env[62552]: DEBUG nova.compute.manager [req-06df3d4c-b354-4cd1-b730-f4a93e073b61 req-c172a45e-9083-4c59-a39a-729418414ab8 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 963.726338] env[62552]: DEBUG nova.compute.manager [req-06df3d4c-b354-4cd1-b730-f4a93e073b61 req-c172a45e-9083-4c59-a39a-729418414ab8 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing instance network info cache due to event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 963.726563] env[62552]: DEBUG oslo_concurrency.lockutils [req-06df3d4c-b354-4cd1-b730-f4a93e073b61 req-c172a45e-9083-4c59-a39a-729418414ab8 service nova] Acquiring lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.726788] env[62552]: DEBUG oslo_concurrency.lockutils [req-06df3d4c-b354-4cd1-b730-f4a93e073b61 req-c172a45e-9083-4c59-a39a-729418414ab8 service nova] Acquired lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.726885] env[62552]: DEBUG nova.network.neutron [req-06df3d4c-b354-4cd1-b730-f4a93e073b61 req-c172a45e-9083-4c59-a39a-729418414ab8 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 963.793791] env[62552]: DEBUG nova.scheduler.client.report [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 963.801480] env[62552]: DEBUG oslo_concurrency.lockutils [None req-34c21a73-b637-40c8-926e-102e53dc2732 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.275s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.939394] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d6fe8c-f9b3-4f41-a80b-d22364e0bf35 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.958366] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance '04186fde-a9ed-415a-bd40-312f0347fcc7' progress to 0 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 964.030869] env[62552]: DEBUG oslo_concurrency.lockutils [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "0418260a-aa27-4955-ab15-b180ec04f0b3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.031178] env[62552]: DEBUG oslo_concurrency.lockutils [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.031368] env[62552]: DEBUG oslo_concurrency.lockutils [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "0418260a-aa27-4955-ab15-b180ec04f0b3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.031594] env[62552]: DEBUG oslo_concurrency.lockutils [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.031777] env[62552]: DEBUG oslo_concurrency.lockutils [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.033867] env[62552]: INFO nova.compute.manager [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Terminating instance [ 964.305885] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.809s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.308866] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.976s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.311944] env[62552]: INFO nova.compute.claims [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.330896] env[62552]: INFO nova.scheduler.client.report [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Deleted allocations for instance c5ed17ee-7421-49d7-975f-1515de1e4b01 [ 964.381385] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.381705] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.465356] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.465684] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd71ff19-e072-445d-ad1d-92908a3ac78c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.474089] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 964.474089] env[62552]: value = "task-1239811" [ 964.474089] env[62552]: _type = "Task" [ 964.474089] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.483822] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239811, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.537494] env[62552]: DEBUG nova.compute.manager [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 964.537756] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.538800] env[62552]: DEBUG nova.network.neutron [req-06df3d4c-b354-4cd1-b730-f4a93e073b61 req-c172a45e-9083-4c59-a39a-729418414ab8 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updated VIF entry in instance network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 964.539169] env[62552]: DEBUG nova.network.neutron [req-06df3d4c-b354-4cd1-b730-f4a93e073b61 req-c172a45e-9083-4c59-a39a-729418414ab8 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.542232] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8b82a8-2c1f-40f5-bc9b-6539d19b87d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.550963] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.550963] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b9867bc-87f7-4964-a6b7-a32f77a0f23e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.557350] env[62552]: DEBUG oslo_vmware.api [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 964.557350] env[62552]: value = "task-1239812" [ 964.557350] env[62552]: _type = "Task" [ 964.557350] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.565449] env[62552]: DEBUG oslo_vmware.api [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.845614] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1e138dc6-0ef7-4476-8694-c4797c5d23eb tempest-ImagesTestJSON-396924259 tempest-ImagesTestJSON-396924259-project-member] Lock "c5ed17ee-7421-49d7-975f-1515de1e4b01" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.628s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.883992] env[62552]: DEBUG nova.compute.manager [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 964.985451] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239811, 'name': PowerOffVM_Task, 'duration_secs': 0.214779} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.985863] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.985999] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance '04186fde-a9ed-415a-bd40-312f0347fcc7' progress to 17 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 965.045425] env[62552]: DEBUG oslo_concurrency.lockutils [req-06df3d4c-b354-4cd1-b730-f4a93e073b61 req-c172a45e-9083-4c59-a39a-729418414ab8 service nova] Releasing lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.067857] env[62552]: DEBUG oslo_vmware.api [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239812, 'name': PowerOffVM_Task, 'duration_secs': 0.279287} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.067857] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.067857] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.068103] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2ee698d-b2db-48a6-93d6-215dd697073b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.127327] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 965.127573] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 965.127762] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleting the datastore file [datastore2] 0418260a-aa27-4955-ab15-b180ec04f0b3 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.128059] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42f6d9ba-1230-4a60-99c5-edeaaa30bd8a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.134238] env[62552]: DEBUG oslo_vmware.api [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 965.134238] env[62552]: value = "task-1239814" [ 965.134238] env[62552]: _type = "Task" [ 965.134238] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.142370] env[62552]: DEBUG oslo_vmware.api [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239814, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.406383] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.494077] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.494077] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.494077] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.494077] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.494077] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.494077] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.494653] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.494960] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.498019] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.498019] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.498019] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.503917] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37e63962-7070-483e-a0ed-ac216fcd323a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.522020] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 965.522020] env[62552]: value = "task-1239815" [ 965.522020] env[62552]: _type = "Task" [ 965.522020] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.531784] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239815, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.563018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b4000e-5557-427c-94a9-a54ce22eb892 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.569479] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9962fcd7-6682-4861-957a-5c72e084e434 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.601766] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a5b999-8eaa-4f33-8ed9-764d43865054 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.609533] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b23ba84-d056-4e5e-a7d8-d96b8c50f9f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.623357] env[62552]: DEBUG nova.compute.provider_tree [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.644384] env[62552]: DEBUG oslo_vmware.api [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239814, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.036162] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239815, 'name': ReconfigVM_Task, 'duration_secs': 0.310362} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.037181] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance '04186fde-a9ed-415a-bd40-312f0347fcc7' progress to 33 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 966.127838] env[62552]: DEBUG nova.scheduler.client.report [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 966.143893] env[62552]: DEBUG oslo_vmware.api [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239814, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.525649} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.144182] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.144373] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.144809] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.144809] env[62552]: INFO nova.compute.manager [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Took 1.61 seconds to destroy the instance on the hypervisor. [ 966.144961] env[62552]: DEBUG oslo.service.loopingcall [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.145783] env[62552]: DEBUG nova.compute.manager [-] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 966.145890] env[62552]: DEBUG nova.network.neutron [-] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 966.543429] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.543642] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.543982] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.544205] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.544337] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.544529] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.544773] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.544988] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.545183] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.545377] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.545567] env[62552]: DEBUG nova.virt.hardware [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.551561] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Reconfiguring VM instance instance-00000054 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 966.551909] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a2c9188-cffd-49f6-99e9-50f7c99f6ecd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.571437] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 966.571437] env[62552]: value = "task-1239816" [ 966.571437] env[62552]: _type = "Task" [ 966.571437] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.579945] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239816, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.633260] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.637839] env[62552]: DEBUG nova.compute.manager [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 966.639483] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.305s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.641131] env[62552]: INFO nova.compute.claims [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.887076] env[62552]: DEBUG nova.compute.manager [req-ce549ff2-16b0-4f3c-9f80-7464f6601b1c req-dbcc9ba6-67e8-4e3b-a93f-25212f467c64 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Received event network-vif-deleted-abe02bcf-41b7-4884-8ef3-70b5268480f0 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 966.887299] env[62552]: INFO nova.compute.manager [req-ce549ff2-16b0-4f3c-9f80-7464f6601b1c req-dbcc9ba6-67e8-4e3b-a93f-25212f467c64 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Neutron deleted interface abe02bcf-41b7-4884-8ef3-70b5268480f0; detaching it from the instance and deleting it from the info cache [ 966.887484] env[62552]: DEBUG nova.network.neutron [req-ce549ff2-16b0-4f3c-9f80-7464f6601b1c req-dbcc9ba6-67e8-4e3b-a93f-25212f467c64 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.081747] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239816, 'name': ReconfigVM_Task, 'duration_secs': 0.173138} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.082363] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Reconfigured VM instance instance-00000054 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 967.083303] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae618ab-1b1d-46b0-888b-78d2d098eb5b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.113743] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 04186fde-a9ed-415a-bd40-312f0347fcc7/04186fde-a9ed-415a-bd40-312f0347fcc7.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.114135] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cb0d872-3586-41b5-bd28-7ab789aa0ee5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.141640] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 967.141640] env[62552]: value = "task-1239817" [ 967.141640] env[62552]: _type = "Task" [ 967.141640] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.146061] env[62552]: DEBUG nova.compute.utils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.150915] env[62552]: DEBUG nova.compute.manager [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 967.151106] env[62552]: DEBUG nova.network.neutron [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 967.161779] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239817, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.198585] env[62552]: DEBUG nova.policy [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb432303c2394f8795672ceaf02f28aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a0bc011d6794602b2bbe1fc01e4c8b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 967.336845] env[62552]: DEBUG nova.network.neutron [-] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.393638] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e9c83bf-4dff-4707-9f48-5a9e33165e8f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.404841] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7b3cc7-1f6a-4861-99e5-347606050696 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.449213] env[62552]: DEBUG nova.compute.manager [req-ce549ff2-16b0-4f3c-9f80-7464f6601b1c req-dbcc9ba6-67e8-4e3b-a93f-25212f467c64 service nova] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Detach interface failed, port_id=abe02bcf-41b7-4884-8ef3-70b5268480f0, reason: Instance 0418260a-aa27-4955-ab15-b180ec04f0b3 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 967.539722] env[62552]: DEBUG nova.network.neutron [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Successfully created port: a1557e27-81a5-4443-a4f2-15be22c30b3c {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.656770] env[62552]: DEBUG nova.compute.manager [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 967.661956] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239817, 'name': ReconfigVM_Task, 'duration_secs': 0.268179} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.662678] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 04186fde-a9ed-415a-bd40-312f0347fcc7/04186fde-a9ed-415a-bd40-312f0347fcc7.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.662942] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance '04186fde-a9ed-415a-bd40-312f0347fcc7' progress to 50 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 967.840259] env[62552]: INFO nova.compute.manager [-] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Took 1.69 seconds to deallocate network for instance. [ 967.902868] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996431e0-fbb1-4eb5-a8ce-2af89d099229 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.911627] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608e053d-ed2c-4dc0-8d30-e46d575b6790 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.944019] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3f0be3-2bcc-4fc6-8a3f-b03fcb308c21 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.954283] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05757dcb-1c1f-4cba-b075-a535bac8ed39 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.971816] env[62552]: DEBUG nova.compute.provider_tree [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 968.172731] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987af3da-2d55-4e5e-851b-f93efd910082 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.178049] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "cdee22a9-4327-47af-ab14-8403a06b2802" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.178159] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.212995] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b71a8c-cb16-4ca4-9a27-393d68097e2d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.234506] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance '04186fde-a9ed-415a-bd40-312f0347fcc7' progress to 67 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 968.349582] env[62552]: DEBUG oslo_concurrency.lockutils [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.501897] env[62552]: ERROR nova.scheduler.client.report [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [req-43ba8d5c-327a-4a2a-98a4-f7e37b2b082d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-43ba8d5c-327a-4a2a-98a4-f7e37b2b082d"}]} [ 968.521592] env[62552]: DEBUG nova.scheduler.client.report [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 968.543418] env[62552]: DEBUG nova.scheduler.client.report [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 968.544714] env[62552]: DEBUG nova.compute.provider_tree [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 968.556906] env[62552]: DEBUG nova.scheduler.client.report [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 968.583982] env[62552]: DEBUG nova.scheduler.client.report [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 968.673140] env[62552]: DEBUG nova.compute.manager [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 968.684424] env[62552]: INFO nova.compute.manager [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Detaching volume 250ad8c9-818e-4ebf-864a-a2e15cd300d3 [ 968.704155] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.704423] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.704939] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.704939] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.705063] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.705310] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.705506] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.705778] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.705910] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.706093] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.706272] env[62552]: DEBUG nova.virt.hardware [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.707273] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1984fc0-5237-4ca2-b465-9ed7f29ae62b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.719089] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18cebf82-7c7f-4579-800e-4f9e8262c9ee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.738373] env[62552]: INFO nova.virt.block_device [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Attempting to driver detach volume 250ad8c9-818e-4ebf-864a-a2e15cd300d3 from mountpoint /dev/sdb [ 968.738373] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 968.738373] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267441', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'name': 'volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cdee22a9-4327-47af-ab14-8403a06b2802', 'attached_at': '', 'detached_at': '', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'serial': '250ad8c9-818e-4ebf-864a-a2e15cd300d3'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 968.738373] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a8c0e7-ee31-4479-acbc-9ada2e67c1e4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.769380] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90c693a-373d-4ec6-8309-6765e38cb1f4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.777757] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2334bdbf-33c4-41ba-91ae-69e8090fd858 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.803009] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f3e087-7e0a-41e5-9170-5266502c7261 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.818049] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] The volume has not been displaced from its original location: [datastore1] volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3/volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 968.823294] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Reconfiguring VM instance instance-0000002d to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 968.826420] env[62552]: DEBUG nova.network.neutron [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Port 04e39cd5-39ae-4671-b715-e300dec6021d binding to destination host cpu-1 is already ACTIVE {{(pid=62552) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 968.827726] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddebf221-94f2-47b7-bfb7-d80a1271bd7e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.851503] env[62552]: DEBUG oslo_vmware.api [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 968.851503] env[62552]: value = "task-1239818" [ 968.851503] env[62552]: _type = "Task" [ 968.851503] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.861771] env[62552]: DEBUG oslo_vmware.api [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239818, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.875000] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2efec816-0360-4cff-91ab-1517cfe0ac68 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.883233] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c6f226-e3e7-449f-a00e-02442b5e3bc7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.916644] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b15c3c-8de8-4f36-b4f0-cde00bd1b35e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.925079] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8adfeff-4031-4566-a2d8-de8a66a6f3b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.939886] env[62552]: DEBUG nova.compute.provider_tree [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 969.192581] env[62552]: DEBUG nova.compute.manager [req-e0531fad-2f9f-4204-ba17-67377f0bab39 req-77e74c70-abaa-4289-bfc7-b3bd173c94d5 service nova] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Received event network-vif-plugged-a1557e27-81a5-4443-a4f2-15be22c30b3c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 969.192869] env[62552]: DEBUG oslo_concurrency.lockutils [req-e0531fad-2f9f-4204-ba17-67377f0bab39 req-77e74c70-abaa-4289-bfc7-b3bd173c94d5 service nova] Acquiring lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.193027] env[62552]: DEBUG oslo_concurrency.lockutils [req-e0531fad-2f9f-4204-ba17-67377f0bab39 req-77e74c70-abaa-4289-bfc7-b3bd173c94d5 service nova] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.193207] env[62552]: DEBUG oslo_concurrency.lockutils [req-e0531fad-2f9f-4204-ba17-67377f0bab39 req-77e74c70-abaa-4289-bfc7-b3bd173c94d5 service nova] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.193478] env[62552]: DEBUG nova.compute.manager [req-e0531fad-2f9f-4204-ba17-67377f0bab39 req-77e74c70-abaa-4289-bfc7-b3bd173c94d5 service nova] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] No waiting events found dispatching network-vif-plugged-a1557e27-81a5-4443-a4f2-15be22c30b3c {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 969.193657] env[62552]: WARNING nova.compute.manager [req-e0531fad-2f9f-4204-ba17-67377f0bab39 req-77e74c70-abaa-4289-bfc7-b3bd173c94d5 service nova] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Received unexpected event network-vif-plugged-a1557e27-81a5-4443-a4f2-15be22c30b3c for instance with vm_state building and task_state spawning. [ 969.277630] env[62552]: DEBUG nova.network.neutron [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Successfully updated port: a1557e27-81a5-4443-a4f2-15be22c30b3c {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.376070] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "04186fde-a9ed-415a-bd40-312f0347fcc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.376070] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.376070] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.381202] env[62552]: DEBUG oslo_vmware.api [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239818, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.476299] env[62552]: DEBUG nova.scheduler.client.report [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 112 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 969.476299] env[62552]: DEBUG nova.compute.provider_tree [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 112 to 113 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 969.476299] env[62552]: DEBUG nova.compute.provider_tree [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 969.785676] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "refresh_cache-ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.785830] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "refresh_cache-ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.785981] env[62552]: DEBUG nova.network.neutron [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.873728] env[62552]: DEBUG oslo_vmware.api [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239818, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.979236] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.340s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.979707] env[62552]: DEBUG nova.compute.manager [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 969.982696] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.320s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.984722] env[62552]: INFO nova.compute.claims [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.377249] env[62552]: DEBUG oslo_vmware.api [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239818, 'name': ReconfigVM_Task, 'duration_secs': 1.268092} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.377249] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Reconfigured VM instance instance-0000002d to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 970.385222] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97d247e0-05fd-4b28-9d09-2a40b81c96c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.398991] env[62552]: DEBUG nova.network.neutron [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 970.403588] env[62552]: DEBUG oslo_vmware.api [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 970.403588] env[62552]: value = "task-1239819" [ 970.403588] env[62552]: _type = "Task" [ 970.403588] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.413839] env[62552]: DEBUG oslo_vmware.api [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239819, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.425648] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.425824] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.426013] env[62552]: DEBUG nova.network.neutron [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.494827] env[62552]: DEBUG nova.compute.utils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.502532] env[62552]: DEBUG nova.compute.manager [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 970.502733] env[62552]: DEBUG nova.network.neutron [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 970.616464] env[62552]: DEBUG nova.policy [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ed5ce401c724f07a3df60d97ac3814b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ecba34d324f42d7af8d1bcf38bbda05', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 970.741422] env[62552]: DEBUG nova.network.neutron [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Updating instance_info_cache with network_info: [{"id": "a1557e27-81a5-4443-a4f2-15be22c30b3c", "address": "fa:16:3e:ec:f7:9a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1557e27-81", "ovs_interfaceid": "a1557e27-81a5-4443-a4f2-15be22c30b3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.915385] env[62552]: DEBUG oslo_vmware.api [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239819, 'name': ReconfigVM_Task, 'duration_secs': 0.172057} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.915995] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267441', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'name': 'volume-250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cdee22a9-4327-47af-ab14-8403a06b2802', 'attached_at': '', 'detached_at': '', 'volume_id': '250ad8c9-818e-4ebf-864a-a2e15cd300d3', 'serial': '250ad8c9-818e-4ebf-864a-a2e15cd300d3'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 971.003756] env[62552]: DEBUG nova.compute.manager [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 971.123037] env[62552]: DEBUG nova.network.neutron [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Successfully created port: 4294380b-fd37-48f3-ba91-6e169312e2a0 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 971.228825] env[62552]: DEBUG nova.compute.manager [req-a9334678-dd93-4d6b-ad3a-2b2849129e29 req-05c6fbd0-cd60-40f4-8e69-6a1f2714c17b service nova] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Received event network-changed-a1557e27-81a5-4443-a4f2-15be22c30b3c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 971.229037] env[62552]: DEBUG nova.compute.manager [req-a9334678-dd93-4d6b-ad3a-2b2849129e29 req-05c6fbd0-cd60-40f4-8e69-6a1f2714c17b service nova] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Refreshing instance network info cache due to event network-changed-a1557e27-81a5-4443-a4f2-15be22c30b3c. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 971.229245] env[62552]: DEBUG oslo_concurrency.lockutils [req-a9334678-dd93-4d6b-ad3a-2b2849129e29 req-05c6fbd0-cd60-40f4-8e69-6a1f2714c17b service nova] Acquiring lock "refresh_cache-ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.244280] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "refresh_cache-ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.244602] env[62552]: DEBUG nova.compute.manager [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Instance network_info: |[{"id": "a1557e27-81a5-4443-a4f2-15be22c30b3c", "address": "fa:16:3e:ec:f7:9a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1557e27-81", "ovs_interfaceid": "a1557e27-81a5-4443-a4f2-15be22c30b3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 971.248706] env[62552]: DEBUG oslo_concurrency.lockutils [req-a9334678-dd93-4d6b-ad3a-2b2849129e29 req-05c6fbd0-cd60-40f4-8e69-6a1f2714c17b service nova] Acquired lock "refresh_cache-ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.248706] env[62552]: DEBUG nova.network.neutron [req-a9334678-dd93-4d6b-ad3a-2b2849129e29 req-05c6fbd0-cd60-40f4-8e69-6a1f2714c17b service nova] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Refreshing network info cache for port a1557e27-81a5-4443-a4f2-15be22c30b3c {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 971.251179] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:f7:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1557e27-81a5-4443-a4f2-15be22c30b3c', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.258324] env[62552]: DEBUG oslo.service.loopingcall [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.262565] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 971.263050] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3cf4fef5-adc3-49af-889e-ef8ea0825667 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.288660] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 971.288660] env[62552]: value = "task-1239820" [ 971.288660] env[62552]: _type = "Task" [ 971.288660] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.297495] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239820, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.319230] env[62552]: DEBUG nova.network.neutron [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance_info_cache with network_info: [{"id": "04e39cd5-39ae-4671-b715-e300dec6021d", "address": "fa:16:3e:6e:65:88", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04e39cd5-39", "ovs_interfaceid": "04e39cd5-39ae-4671-b715-e300dec6021d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.327724] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2a6c46-da6b-4352-a96e-f0cb41544f21 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.335639] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fe605d-0252-489a-9062-90d6888314db {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.368180] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76aa767e-eac9-4245-ac41-f5708f33ff06 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.376014] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6042ff9f-b930-4a40-91d4-800720b8302d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.393095] env[62552]: DEBUG nova.compute.provider_tree [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.481066] env[62552]: DEBUG nova.objects.instance [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'flavor' on Instance uuid cdee22a9-4327-47af-ab14-8403a06b2802 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.801654] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239820, 'name': CreateVM_Task, 'duration_secs': 0.334709} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.804300] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 971.806527] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.806527] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.806527] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 971.807201] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75786afb-dfab-46f2-ba54-55b761b73c4f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.812939] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 971.812939] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a57b0b-7a29-b67c-bfc5-dead21b5131b" [ 971.812939] env[62552]: _type = "Task" [ 971.812939] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.822753] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.827666] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a57b0b-7a29-b67c-bfc5-dead21b5131b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.896886] env[62552]: DEBUG nova.scheduler.client.report [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 972.022571] env[62552]: DEBUG nova.compute.manager [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 972.055568] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 972.056393] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 972.056548] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 972.058566] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 972.058566] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 972.058566] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 972.058566] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 972.058566] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 972.058566] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 972.058566] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 972.058566] env[62552]: DEBUG nova.virt.hardware [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.059086] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d31e05-caba-496a-9597-9cf5d41f1824 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.070973] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54aa6094-6fa5-4936-9a27-5b1a4f3c5a0d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.131017] env[62552]: DEBUG nova.network.neutron [req-a9334678-dd93-4d6b-ad3a-2b2849129e29 req-05c6fbd0-cd60-40f4-8e69-6a1f2714c17b service nova] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Updated VIF entry in instance network info cache for port a1557e27-81a5-4443-a4f2-15be22c30b3c. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 972.131017] env[62552]: DEBUG nova.network.neutron [req-a9334678-dd93-4d6b-ad3a-2b2849129e29 req-05c6fbd0-cd60-40f4-8e69-6a1f2714c17b service nova] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Updating instance_info_cache with network_info: [{"id": "a1557e27-81a5-4443-a4f2-15be22c30b3c", "address": "fa:16:3e:ec:f7:9a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1557e27-81", "ovs_interfaceid": "a1557e27-81a5-4443-a4f2-15be22c30b3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.324761] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a57b0b-7a29-b67c-bfc5-dead21b5131b, 'name': SearchDatastore_Task, 'duration_secs': 0.008952} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.325340] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.325340] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.325569] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.325712] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.325893] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.326290] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4250de7-1f6a-46c0-aa4f-51ca1fbcc611 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.337270] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.337270] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 972.339676] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e625c0f-5022-4dcc-bfbe-1fbe234eefe7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.346241] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 972.346241] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527201de-942a-0be4-a0c3-5eb299813981" [ 972.346241] env[62552]: _type = "Task" [ 972.346241] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.355220] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527201de-942a-0be4-a0c3-5eb299813981, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.356669] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1f23f5-a1d0-4010-944f-2f7f66ffbe03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.375339] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fc579d-70f0-4fe2-a942-89d27f20408f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.383260] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance '04186fde-a9ed-415a-bd40-312f0347fcc7' progress to 83 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 972.406633] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.406968] env[62552]: DEBUG nova.compute.manager [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 972.410873] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.448s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.412628] env[62552]: INFO nova.compute.claims [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.491521] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a56a9ff8-ee0e-4d59-938d-f7d105015944 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.313s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.631796] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "cdee22a9-4327-47af-ab14-8403a06b2802" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.632076] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.632295] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.632485] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.632699] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.634739] env[62552]: DEBUG oslo_concurrency.lockutils [req-a9334678-dd93-4d6b-ad3a-2b2849129e29 req-05c6fbd0-cd60-40f4-8e69-6a1f2714c17b service nova] Releasing lock "refresh_cache-ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.636625] env[62552]: INFO nova.compute.manager [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Terminating instance [ 972.859293] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527201de-942a-0be4-a0c3-5eb299813981, 'name': SearchDatastore_Task, 'duration_secs': 0.009596} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.860119] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83ee206a-6098-43df-aab2-dac54a8fd807 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.866105] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 972.866105] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524a7d57-3590-1469-e85a-9970987f5eb7" [ 972.866105] env[62552]: _type = "Task" [ 972.866105] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.874192] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524a7d57-3590-1469-e85a-9970987f5eb7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.889981] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.890270] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b6ce519-d30e-45a7-824e-eb87a1f3d218 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.899280] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 972.899280] env[62552]: value = "task-1239821" [ 972.899280] env[62552]: _type = "Task" [ 972.899280] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.908901] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239821, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.915017] env[62552]: DEBUG nova.compute.utils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 972.915345] env[62552]: DEBUG nova.compute.manager [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 972.915563] env[62552]: DEBUG nova.network.neutron [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 973.010801] env[62552]: DEBUG nova.policy [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7f8aa8200874dddb71d8b21bd12ca04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8b8d96b464a439e9c7ef6f3e419a9bc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.141676] env[62552]: DEBUG nova.compute.manager [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 973.142939] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 973.143953] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fc565c-28c0-412b-a469-b1bac218a84c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.151950] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.152436] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46c66573-da3d-467c-beae-79dfde4e1c0f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.159168] env[62552]: DEBUG oslo_vmware.api [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 973.159168] env[62552]: value = "task-1239822" [ 973.159168] env[62552]: _type = "Task" [ 973.159168] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.168361] env[62552]: DEBUG oslo_vmware.api [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.225360] env[62552]: DEBUG nova.compute.manager [req-12023ce1-7a8d-46ee-b8ad-8e4b74e82484 req-0e822d91-6b7d-4333-b9eb-f88e50a7a815 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Received event network-vif-plugged-4294380b-fd37-48f3-ba91-6e169312e2a0 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 973.225838] env[62552]: DEBUG oslo_concurrency.lockutils [req-12023ce1-7a8d-46ee-b8ad-8e4b74e82484 req-0e822d91-6b7d-4333-b9eb-f88e50a7a815 service nova] Acquiring lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.226493] env[62552]: DEBUG oslo_concurrency.lockutils [req-12023ce1-7a8d-46ee-b8ad-8e4b74e82484 req-0e822d91-6b7d-4333-b9eb-f88e50a7a815 service nova] Lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.226771] env[62552]: DEBUG oslo_concurrency.lockutils [req-12023ce1-7a8d-46ee-b8ad-8e4b74e82484 req-0e822d91-6b7d-4333-b9eb-f88e50a7a815 service nova] Lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.227774] env[62552]: DEBUG nova.compute.manager [req-12023ce1-7a8d-46ee-b8ad-8e4b74e82484 req-0e822d91-6b7d-4333-b9eb-f88e50a7a815 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] No waiting events found dispatching network-vif-plugged-4294380b-fd37-48f3-ba91-6e169312e2a0 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 973.228200] env[62552]: WARNING nova.compute.manager [req-12023ce1-7a8d-46ee-b8ad-8e4b74e82484 req-0e822d91-6b7d-4333-b9eb-f88e50a7a815 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Received unexpected event network-vif-plugged-4294380b-fd37-48f3-ba91-6e169312e2a0 for instance with vm_state building and task_state spawning. [ 973.275314] env[62552]: DEBUG nova.network.neutron [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Successfully updated port: 4294380b-fd37-48f3-ba91-6e169312e2a0 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 973.317176] env[62552]: DEBUG nova.network.neutron [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Successfully created port: 54ef4be6-f8fc-458e-a1e0-96a8529e448e {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.383095] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524a7d57-3590-1469-e85a-9970987f5eb7, 'name': SearchDatastore_Task, 'duration_secs': 0.010552} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.384112] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.384112] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71/ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.384222] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c12358b-2ce5-47a3-b568-41b968dc18cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.391095] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 973.391095] env[62552]: value = "task-1239823" [ 973.391095] env[62552]: _type = "Task" [ 973.391095] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.399190] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239823, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.408183] env[62552]: DEBUG oslo_vmware.api [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239821, 'name': PowerOnVM_Task, 'duration_secs': 0.39831} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.408455] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.408631] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c0334dc2-54e2-4120-8610-3e62a114509c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance '04186fde-a9ed-415a-bd40-312f0347fcc7' progress to 100 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 973.418810] env[62552]: DEBUG nova.compute.manager [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 973.499341] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.499754] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.573433] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquiring lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.573688] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.669853] env[62552]: DEBUG oslo_vmware.api [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239822, 'name': PowerOffVM_Task, 'duration_secs': 0.212126} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.672750] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.672964] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 973.673469] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-15f652d8-ce23-4ea3-ab10-42587abcc65e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.713021] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a686956e-f2ca-4993-a6a2-ba61a1c101d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.721387] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654d259b-9fb4-48f7-9fb1-eebbc01f04c4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.759127] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c296c9d-6fea-47a3-8234-7c64aed612ee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.761989] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 973.762264] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 973.762582] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleting the datastore file [datastore1] cdee22a9-4327-47af-ab14-8403a06b2802 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.762714] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42c163b0-2225-432d-81b7-53894833a2ad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.770729] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501ce126-711a-44a9-aa20-04f2ab8abe12 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.774854] env[62552]: DEBUG oslo_vmware.api [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 973.774854] env[62552]: value = "task-1239825" [ 973.774854] env[62552]: _type = "Task" [ 973.774854] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.788231] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquiring lock "refresh_cache-2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.788405] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquired lock "refresh_cache-2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.788524] env[62552]: DEBUG nova.network.neutron [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 973.790450] env[62552]: DEBUG nova.compute.provider_tree [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.797600] env[62552]: DEBUG oslo_vmware.api [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239825, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.901621] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239823, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476419} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.901923] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71/ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 973.902352] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 973.902577] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3fdcc48b-9249-427d-aa31-13acdd1e5da5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.909209] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 973.909209] env[62552]: value = "task-1239826" [ 973.909209] env[62552]: _type = "Task" [ 973.909209] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.923729] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239826, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.004938] env[62552]: DEBUG nova.compute.manager [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 974.076695] env[62552]: DEBUG nova.compute.manager [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 974.285227] env[62552]: DEBUG oslo_vmware.api [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239825, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209673} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.285572] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.285770] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 974.285951] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 974.286147] env[62552]: INFO nova.compute.manager [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Took 1.14 seconds to destroy the instance on the hypervisor. [ 974.286412] env[62552]: DEBUG oslo.service.loopingcall [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.286616] env[62552]: DEBUG nova.compute.manager [-] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 974.286851] env[62552]: DEBUG nova.network.neutron [-] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 974.293458] env[62552]: DEBUG nova.scheduler.client.report [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 974.379416] env[62552]: DEBUG nova.network.neutron [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 974.423098] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239826, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.22539} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.430467] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.430751] env[62552]: DEBUG nova.compute.manager [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 974.433971] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9a8bd6-5c7e-45d5-92a9-0fa254fd6eb8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.458384] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71/ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.460661] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8d62670-9375-4998-b54f-fa937b0819f7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.484019] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 974.484019] env[62552]: value = "task-1239827" [ 974.484019] env[62552]: _type = "Task" [ 974.484019] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.488723] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.489236] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.489572] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.489907] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.490190] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.490571] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.490912] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.492018] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.492018] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.492018] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.492018] env[62552]: DEBUG nova.virt.hardware [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.493409] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485a30fa-b578-47f4-9dfc-b62a3c1bccb8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.505076] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.508063] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ca660f-5020-4448-b427-b8fbe44b4226 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.543286] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.599276] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.798257] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.798795] env[62552]: DEBUG nova.compute.manager [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 974.801960] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.396s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.803524] env[62552]: INFO nova.compute.claims [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.865566] env[62552]: DEBUG nova.network.neutron [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Successfully updated port: 54ef4be6-f8fc-458e-a1e0-96a8529e448e {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 974.998261] env[62552]: DEBUG nova.compute.manager [req-bdad5e23-9039-46db-b792-74df56f52cde req-bcfc6ab3-4677-4b45-b158-a374c7303089 service nova] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Received event network-vif-plugged-54ef4be6-f8fc-458e-a1e0-96a8529e448e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 974.998261] env[62552]: DEBUG oslo_concurrency.lockutils [req-bdad5e23-9039-46db-b792-74df56f52cde req-bcfc6ab3-4677-4b45-b158-a374c7303089 service nova] Acquiring lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.998261] env[62552]: DEBUG oslo_concurrency.lockutils [req-bdad5e23-9039-46db-b792-74df56f52cde req-bcfc6ab3-4677-4b45-b158-a374c7303089 service nova] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.998261] env[62552]: DEBUG oslo_concurrency.lockutils [req-bdad5e23-9039-46db-b792-74df56f52cde req-bcfc6ab3-4677-4b45-b158-a374c7303089 service nova] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.998261] env[62552]: DEBUG nova.compute.manager [req-bdad5e23-9039-46db-b792-74df56f52cde req-bcfc6ab3-4677-4b45-b158-a374c7303089 service nova] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] No waiting events found dispatching network-vif-plugged-54ef4be6-f8fc-458e-a1e0-96a8529e448e {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 974.998261] env[62552]: WARNING nova.compute.manager [req-bdad5e23-9039-46db-b792-74df56f52cde req-bcfc6ab3-4677-4b45-b158-a374c7303089 service nova] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Received unexpected event network-vif-plugged-54ef4be6-f8fc-458e-a1e0-96a8529e448e for instance with vm_state building and task_state spawning. [ 975.000372] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239827, 'name': ReconfigVM_Task, 'duration_secs': 0.305666} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.001329] env[62552]: DEBUG nova.network.neutron [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Updating instance_info_cache with network_info: [{"id": "4294380b-fd37-48f3-ba91-6e169312e2a0", "address": "fa:16:3e:b6:87:b0", "network": {"id": "a4d7c75f-bdf3-41a6-9c2f-347202c7bae9", "bridge": "br-int", "label": "tempest-ServersTestJSON-464632350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecba34d324f42d7af8d1bcf38bbda05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89f807d9-140f-4a6f-8bce-96795f9482ee", "external-id": "nsx-vlan-transportzone-762", "segmentation_id": 762, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4294380b-fd", "ovs_interfaceid": "4294380b-fd37-48f3-ba91-6e169312e2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.002517] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Reconfigured VM instance instance-00000057 to attach disk [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71/ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.005805] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecea01b0-2bf4-49d4-bc10-ac47bf7881e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.013104] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 975.013104] env[62552]: value = "task-1239828" [ 975.013104] env[62552]: _type = "Task" [ 975.013104] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.020838] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239828, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.257589] env[62552]: DEBUG nova.compute.manager [req-8247f05d-32e2-407f-ad55-e3d51ac0014d req-13669642-4649-45a9-b143-dcf114175cd4 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Received event network-changed-4294380b-fd37-48f3-ba91-6e169312e2a0 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 975.257798] env[62552]: DEBUG nova.compute.manager [req-8247f05d-32e2-407f-ad55-e3d51ac0014d req-13669642-4649-45a9-b143-dcf114175cd4 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Refreshing instance network info cache due to event network-changed-4294380b-fd37-48f3-ba91-6e169312e2a0. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 975.257990] env[62552]: DEBUG oslo_concurrency.lockutils [req-8247f05d-32e2-407f-ad55-e3d51ac0014d req-13669642-4649-45a9-b143-dcf114175cd4 service nova] Acquiring lock "refresh_cache-2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.304101] env[62552]: DEBUG nova.compute.utils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.305861] env[62552]: DEBUG nova.compute.manager [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 975.306233] env[62552]: DEBUG nova.network.neutron [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 975.368065] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-304ad9f2-1965-4a70-aec9-0d2c931b8d95" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.368232] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-304ad9f2-1965-4a70-aec9-0d2c931b8d95" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.368448] env[62552]: DEBUG nova.network.neutron [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.417728] env[62552]: DEBUG nova.policy [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bdea5d6f50634b5e89852744db56db67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8495bac65967441996ecca1fd22a4da4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 975.508332] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Releasing lock "refresh_cache-2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.508332] env[62552]: DEBUG nova.compute.manager [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Instance network_info: |[{"id": "4294380b-fd37-48f3-ba91-6e169312e2a0", "address": "fa:16:3e:b6:87:b0", "network": {"id": "a4d7c75f-bdf3-41a6-9c2f-347202c7bae9", "bridge": "br-int", "label": "tempest-ServersTestJSON-464632350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecba34d324f42d7af8d1bcf38bbda05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89f807d9-140f-4a6f-8bce-96795f9482ee", "external-id": "nsx-vlan-transportzone-762", "segmentation_id": 762, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4294380b-fd", "ovs_interfaceid": "4294380b-fd37-48f3-ba91-6e169312e2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 975.508332] env[62552]: DEBUG oslo_concurrency.lockutils [req-8247f05d-32e2-407f-ad55-e3d51ac0014d req-13669642-4649-45a9-b143-dcf114175cd4 service nova] Acquired lock "refresh_cache-2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.508332] env[62552]: DEBUG nova.network.neutron [req-8247f05d-32e2-407f-ad55-e3d51ac0014d req-13669642-4649-45a9-b143-dcf114175cd4 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Refreshing network info cache for port 4294380b-fd37-48f3-ba91-6e169312e2a0 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 975.508332] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:87:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89f807d9-140f-4a6f-8bce-96795f9482ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4294380b-fd37-48f3-ba91-6e169312e2a0', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 975.518931] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Creating folder: Project (9ecba34d324f42d7af8d1bcf38bbda05). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 975.520587] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ef1574c8-d433-4e0b-a4ff-fc0f74ea363c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.534283] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239828, 'name': Rename_Task, 'duration_secs': 0.18821} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.534738] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.535145] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ae2cbdf-b957-434b-82c9-a52c7a3a5b9d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.540966] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Created folder: Project (9ecba34d324f42d7af8d1bcf38bbda05) in parent group-v267339. [ 975.540966] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Creating folder: Instances. Parent ref: group-v267485. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 975.540966] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3962e2c2-47ac-4cc2-be08-cb7735aaead2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.545019] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 975.545019] env[62552]: value = "task-1239830" [ 975.545019] env[62552]: _type = "Task" [ 975.545019] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.553371] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Created folder: Instances in parent group-v267485. [ 975.553371] env[62552]: DEBUG oslo.service.loopingcall [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.553371] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 975.553371] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18c9e34b-98f9-473b-a131-b218ec7251fb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.571676] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239830, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.576510] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 975.576510] env[62552]: value = "task-1239832" [ 975.576510] env[62552]: _type = "Task" [ 975.576510] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.584435] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239832, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.754282] env[62552]: DEBUG nova.network.neutron [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Successfully created port: 7734d400-5df3-4f55-b6da-b52dfd148993 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.811702] env[62552]: DEBUG nova.compute.manager [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 975.909983] env[62552]: DEBUG nova.network.neutron [-] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.928970] env[62552]: DEBUG nova.network.neutron [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 976.053869] env[62552]: DEBUG oslo_vmware.api [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239830, 'name': PowerOnVM_Task, 'duration_secs': 0.504333} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.057280] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 976.057571] env[62552]: INFO nova.compute.manager [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Took 7.38 seconds to spawn the instance on the hypervisor. [ 976.057814] env[62552]: DEBUG nova.compute.manager [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 976.062256] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a8ab96-e56d-4c36-b180-1385b16632f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.087595] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239832, 'name': CreateVM_Task, 'duration_secs': 0.435073} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.088028] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 976.088521] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.088733] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.089139] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 976.089426] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e88f2599-89dd-4557-9050-7721ca9fd96a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.096048] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 976.096048] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5232e8c3-9137-e14a-2ea0-0f9f764ad224" [ 976.096048] env[62552]: _type = "Task" [ 976.096048] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.104760] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5232e8c3-9137-e14a-2ea0-0f9f764ad224, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.153384] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9eb7eff-7f47-4f16-814c-8a5b077bc1fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.161744] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b4c900-ea06-45bb-922a-55145bc46289 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.201975] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b417260e-66e2-49e8-ba6f-52cb253bff99 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.211023] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ff5d54-a8db-4b31-bbbb-1cd04740ea87 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.226454] env[62552]: DEBUG nova.compute.provider_tree [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.254849] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "04186fde-a9ed-415a-bd40-312f0347fcc7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.255124] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.255314] env[62552]: DEBUG nova.compute.manager [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Going to confirm migration 2 {{(pid=62552) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 976.278565] env[62552]: DEBUG nova.network.neutron [req-8247f05d-32e2-407f-ad55-e3d51ac0014d req-13669642-4649-45a9-b143-dcf114175cd4 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Updated VIF entry in instance network info cache for port 4294380b-fd37-48f3-ba91-6e169312e2a0. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 976.279111] env[62552]: DEBUG nova.network.neutron [req-8247f05d-32e2-407f-ad55-e3d51ac0014d req-13669642-4649-45a9-b143-dcf114175cd4 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Updating instance_info_cache with network_info: [{"id": "4294380b-fd37-48f3-ba91-6e169312e2a0", "address": "fa:16:3e:b6:87:b0", "network": {"id": "a4d7c75f-bdf3-41a6-9c2f-347202c7bae9", "bridge": "br-int", "label": "tempest-ServersTestJSON-464632350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecba34d324f42d7af8d1bcf38bbda05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89f807d9-140f-4a6f-8bce-96795f9482ee", "external-id": "nsx-vlan-transportzone-762", "segmentation_id": 762, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4294380b-fd", "ovs_interfaceid": "4294380b-fd37-48f3-ba91-6e169312e2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.295919] env[62552]: DEBUG nova.network.neutron [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Updating instance_info_cache with network_info: [{"id": "54ef4be6-f8fc-458e-a1e0-96a8529e448e", "address": "fa:16:3e:0c:fc:7a", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54ef4be6-f8", "ovs_interfaceid": "54ef4be6-f8fc-458e-a1e0-96a8529e448e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.416555] env[62552]: INFO nova.compute.manager [-] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Took 2.13 seconds to deallocate network for instance. [ 976.582815] env[62552]: INFO nova.compute.manager [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Took 20.27 seconds to build instance. [ 976.608170] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5232e8c3-9137-e14a-2ea0-0f9f764ad224, 'name': SearchDatastore_Task, 'duration_secs': 0.018796} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.608428] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.608666] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.608933] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.609107] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.609296] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.609554] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-130203e1-d698-4309-a4c3-344cf3c5df12 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.624130] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.624320] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.625039] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4288c01-24ce-461a-956a-6fb795f1926f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.629957] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 976.629957] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522c9285-0e0d-8621-a3c7-504381d200d6" [ 976.629957] env[62552]: _type = "Task" [ 976.629957] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.637574] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522c9285-0e0d-8621-a3c7-504381d200d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.733435] env[62552]: DEBUG nova.scheduler.client.report [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 976.783123] env[62552]: DEBUG oslo_concurrency.lockutils [req-8247f05d-32e2-407f-ad55-e3d51ac0014d req-13669642-4649-45a9-b143-dcf114175cd4 service nova] Releasing lock "refresh_cache-2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.791828] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.792022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.792208] env[62552]: DEBUG nova.network.neutron [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 976.792393] env[62552]: DEBUG nova.objects.instance [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lazy-loading 'info_cache' on Instance uuid 04186fde-a9ed-415a-bd40-312f0347fcc7 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.798225] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-304ad9f2-1965-4a70-aec9-0d2c931b8d95" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.798486] env[62552]: DEBUG nova.compute.manager [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Instance network_info: |[{"id": "54ef4be6-f8fc-458e-a1e0-96a8529e448e", "address": "fa:16:3e:0c:fc:7a", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54ef4be6-f8", "ovs_interfaceid": "54ef4be6-f8fc-458e-a1e0-96a8529e448e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 976.798878] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:fc:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '54ef4be6-f8fc-458e-a1e0-96a8529e448e', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 976.806239] env[62552]: DEBUG oslo.service.loopingcall [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.806940] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 976.807188] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a07f4236-2f7d-4e8f-97bd-fdb69d8a2294 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.824153] env[62552]: DEBUG nova.compute.manager [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 976.827619] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 976.827619] env[62552]: value = "task-1239833" [ 976.827619] env[62552]: _type = "Task" [ 976.827619] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.839275] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239833, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.847695] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.847937] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.848112] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.848327] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.848524] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.848750] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.848990] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.849175] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.849351] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.849521] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.849701] env[62552]: DEBUG nova.virt.hardware [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.850505] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3324a0d9-20cf-44f4-94c2-3b78842b7162 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.859748] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42e016a-f97d-4f1e-aeef-02acada540fd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.922116] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.022678] env[62552]: DEBUG nova.compute.manager [req-ed1e8cbc-8d92-4de1-b8d4-1e5c2632a2f6 req-fcfb0f08-ead3-4fe8-93da-5f6545556e58 service nova] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Received event network-changed-54ef4be6-f8fc-458e-a1e0-96a8529e448e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 977.022849] env[62552]: DEBUG nova.compute.manager [req-ed1e8cbc-8d92-4de1-b8d4-1e5c2632a2f6 req-fcfb0f08-ead3-4fe8-93da-5f6545556e58 service nova] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Refreshing instance network info cache due to event network-changed-54ef4be6-f8fc-458e-a1e0-96a8529e448e. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 977.023093] env[62552]: DEBUG oslo_concurrency.lockutils [req-ed1e8cbc-8d92-4de1-b8d4-1e5c2632a2f6 req-fcfb0f08-ead3-4fe8-93da-5f6545556e58 service nova] Acquiring lock "refresh_cache-304ad9f2-1965-4a70-aec9-0d2c931b8d95" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.023267] env[62552]: DEBUG oslo_concurrency.lockutils [req-ed1e8cbc-8d92-4de1-b8d4-1e5c2632a2f6 req-fcfb0f08-ead3-4fe8-93da-5f6545556e58 service nova] Acquired lock "refresh_cache-304ad9f2-1965-4a70-aec9-0d2c931b8d95" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.023446] env[62552]: DEBUG nova.network.neutron [req-ed1e8cbc-8d92-4de1-b8d4-1e5c2632a2f6 req-fcfb0f08-ead3-4fe8-93da-5f6545556e58 service nova] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Refreshing network info cache for port 54ef4be6-f8fc-458e-a1e0-96a8529e448e {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.085193] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6e1c179-92c7-477b-80b6-425f5d8f8c5b tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.779s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.140018] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]522c9285-0e0d-8621-a3c7-504381d200d6, 'name': SearchDatastore_Task, 'duration_secs': 0.042658} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.140835] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a446e88a-34f6-4862-97e4-769a0a3d07c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.146413] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 977.146413] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520319fb-8baf-0e3b-c584-9f11babf6897" [ 977.146413] env[62552]: _type = "Task" [ 977.146413] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.154832] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520319fb-8baf-0e3b-c584-9f11babf6897, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.238926] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.239727] env[62552]: DEBUG nova.compute.manager [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 977.243989] env[62552]: DEBUG oslo_concurrency.lockutils [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.894s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.244366] env[62552]: DEBUG nova.objects.instance [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'resources' on Instance uuid 0418260a-aa27-4955-ab15-b180ec04f0b3 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.283539] env[62552]: DEBUG nova.compute.manager [req-16c8208a-2ea5-4e86-aa6c-977f608679e0 req-356d6405-97ba-4f1d-895c-78ade8c4b21a service nova] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Received event network-vif-deleted-929d55b0-5a35-4d2d-a172-15d79a215977 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 977.340742] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239833, 'name': CreateVM_Task, 'duration_secs': 0.366835} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.340994] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 977.341657] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.341879] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.342189] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.342453] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-387509c6-d28b-4e51-b4b5-928adbc87d76 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.347635] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 977.347635] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520e4365-31f8-ce1e-c474-916245d9c5f1" [ 977.347635] env[62552]: _type = "Task" [ 977.347635] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.356171] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520e4365-31f8-ce1e-c474-916245d9c5f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.660022] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520319fb-8baf-0e3b-c584-9f11babf6897, 'name': SearchDatastore_Task, 'duration_secs': 0.012475} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.660022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.660022] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734/2c8a5f30-b54f-42ff-a5e0-04eceb1ac734.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.660022] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-48422243-d6ed-4c66-880f-1774733cbba3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.665615] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 977.665615] env[62552]: value = "task-1239834" [ 977.665615] env[62552]: _type = "Task" [ 977.665615] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.675674] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.747881] env[62552]: DEBUG nova.compute.utils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.754093] env[62552]: DEBUG nova.compute.manager [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 977.754673] env[62552]: DEBUG nova.network.neutron [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 977.814876] env[62552]: DEBUG nova.policy [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6515846d33e44bca942767f9e1d650b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2f57f72de0c4adc9754458f4e7040b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 977.867112] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520e4365-31f8-ce1e-c474-916245d9c5f1, 'name': SearchDatastore_Task, 'duration_secs': 0.013425} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.867584] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.868108] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.868416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.868620] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.868963] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.869180] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84057657-b4cf-4164-92eb-0a50464884c1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.890407] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.890607] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 977.891602] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-001a42c7-cc92-4175-b1b3-9f3d4966bf80 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.901549] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 977.901549] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5217a39a-5308-dde8-5617-f2b1b7ac0103" [ 977.901549] env[62552]: _type = "Task" [ 977.901549] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.917385] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5217a39a-5308-dde8-5617-f2b1b7ac0103, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.056246] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fc3060-4d7b-41bd-a613-1992e7cbf68c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.065710] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5058921a-2ae1-457c-8b15-3d1e5fd3ae7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.103979] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741da8e5-5c43-4020-a78c-bc0b8b652dfe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.114027] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70383e84-3e7b-4431-9bae-ffd3f66324b2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.128992] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.129373] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.129613] env[62552]: DEBUG nova.compute.manager [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 978.130156] env[62552]: DEBUG nova.compute.provider_tree [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.131946] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60a14fa-a358-4952-8b5a-632cb03e5370 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.138116] env[62552]: DEBUG nova.compute.manager [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62552) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 978.138687] env[62552]: DEBUG nova.objects.instance [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lazy-loading 'flavor' on Instance uuid ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.155926] env[62552]: DEBUG nova.network.neutron [req-ed1e8cbc-8d92-4de1-b8d4-1e5c2632a2f6 req-fcfb0f08-ead3-4fe8-93da-5f6545556e58 service nova] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Updated VIF entry in instance network info cache for port 54ef4be6-f8fc-458e-a1e0-96a8529e448e. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.156312] env[62552]: DEBUG nova.network.neutron [req-ed1e8cbc-8d92-4de1-b8d4-1e5c2632a2f6 req-fcfb0f08-ead3-4fe8-93da-5f6545556e58 service nova] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Updating instance_info_cache with network_info: [{"id": "54ef4be6-f8fc-458e-a1e0-96a8529e448e", "address": "fa:16:3e:0c:fc:7a", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54ef4be6-f8", "ovs_interfaceid": "54ef4be6-f8fc-458e-a1e0-96a8529e448e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.176585] env[62552]: DEBUG nova.network.neutron [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Successfully updated port: 7734d400-5df3-4f55-b6da-b52dfd148993 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 978.178169] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239834, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496214} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.178688] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734/2c8a5f30-b54f-42ff-a5e0-04eceb1ac734.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 978.178946] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 978.179250] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf18f1ed-412e-4693-a565-315f68c1bab2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.183348] env[62552]: DEBUG nova.network.neutron [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Successfully created port: 34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.190318] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 978.190318] env[62552]: value = "task-1239835" [ 978.190318] env[62552]: _type = "Task" [ 978.190318] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.201384] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239835, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.254557] env[62552]: DEBUG nova.compute.manager [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 978.292733] env[62552]: DEBUG nova.network.neutron [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance_info_cache with network_info: [{"id": "04e39cd5-39ae-4671-b715-e300dec6021d", "address": "fa:16:3e:6e:65:88", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04e39cd5-39", "ovs_interfaceid": "04e39cd5-39ae-4671-b715-e300dec6021d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.412591] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5217a39a-5308-dde8-5617-f2b1b7ac0103, 'name': SearchDatastore_Task, 'duration_secs': 0.056871} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.415427] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b286a597-8b92-41ef-8ac3-150e6509573a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.420804] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 978.420804] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52961662-8261-1728-2745-6e3b59ee007a" [ 978.420804] env[62552]: _type = "Task" [ 978.420804] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.432024] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52961662-8261-1728-2745-6e3b59ee007a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.636376] env[62552]: DEBUG nova.scheduler.client.report [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 978.659120] env[62552]: DEBUG oslo_concurrency.lockutils [req-ed1e8cbc-8d92-4de1-b8d4-1e5c2632a2f6 req-fcfb0f08-ead3-4fe8-93da-5f6545556e58 service nova] Releasing lock "refresh_cache-304ad9f2-1965-4a70-aec9-0d2c931b8d95" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.679602] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-505d7f8e-c27f-487e-98d8-c840a526d40e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.679923] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-505d7f8e-c27f-487e-98d8-c840a526d40e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.679981] env[62552]: DEBUG nova.network.neutron [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 978.700295] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239835, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065116} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.700646] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.701440] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7bc491-7d49-475f-8f7d-dd716cd6e851 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.723590] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734/2c8a5f30-b54f-42ff-a5e0-04eceb1ac734.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.724281] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eca32dc4-f56c-4336-94da-486322315004 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.745333] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 978.745333] env[62552]: value = "task-1239836" [ 978.745333] env[62552]: _type = "Task" [ 978.745333] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.753329] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239836, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.794447] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-04186fde-a9ed-415a-bd40-312f0347fcc7" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.794731] env[62552]: DEBUG nova.objects.instance [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lazy-loading 'migration_context' on Instance uuid 04186fde-a9ed-415a-bd40-312f0347fcc7 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.931476] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52961662-8261-1728-2745-6e3b59ee007a, 'name': SearchDatastore_Task, 'duration_secs': 0.033277} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.932175] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.932443] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 304ad9f2-1965-4a70-aec9-0d2c931b8d95/304ad9f2-1965-4a70-aec9-0d2c931b8d95.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 978.932734] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6197780-087e-4946-ab43-8f8a3c07eb6b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.939655] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 978.939655] env[62552]: value = "task-1239837" [ 978.939655] env[62552]: _type = "Task" [ 978.939655] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.947397] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239837, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.141967] env[62552]: DEBUG oslo_concurrency.lockutils [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.898s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.145264] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.601s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.146834] env[62552]: INFO nova.compute.claims [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.149404] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.149925] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31413506-f041-49cc-9ae0-3e4761d2c062 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.157828] env[62552]: DEBUG oslo_vmware.api [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 979.157828] env[62552]: value = "task-1239838" [ 979.157828] env[62552]: _type = "Task" [ 979.157828] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.167649] env[62552]: DEBUG oslo_vmware.api [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239838, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.172386] env[62552]: INFO nova.scheduler.client.report [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleted allocations for instance 0418260a-aa27-4955-ab15-b180ec04f0b3 [ 979.232966] env[62552]: DEBUG nova.network.neutron [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 979.257724] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239836, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.263607] env[62552]: DEBUG nova.compute.manager [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 979.296109] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.296382] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.296544] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.296734] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.296885] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.297144] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.297265] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.297416] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.297648] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.297786] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.297997] env[62552]: DEBUG nova.virt.hardware [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.298840] env[62552]: DEBUG nova.objects.base [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Object Instance<04186fde-a9ed-415a-bd40-312f0347fcc7> lazy-loaded attributes: info_cache,migration_context {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 979.299728] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37170a3-28e2-42f9-bfc0-ec6c4471be75 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.305575] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6578b82b-04e7-4248-bd3e-ff8c78a4b6b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.329341] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534b792a-4e8e-4c64-bc79-3c25fa6870bb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.336702] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89afae9e-f703-48d2-9b41-fbeb2669d727 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.342946] env[62552]: DEBUG nova.compute.manager [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Received event network-vif-plugged-7734d400-5df3-4f55-b6da-b52dfd148993 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 979.343170] env[62552]: DEBUG oslo_concurrency.lockutils [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] Acquiring lock "505d7f8e-c27f-487e-98d8-c840a526d40e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.343421] env[62552]: DEBUG oslo_concurrency.lockutils [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] Lock "505d7f8e-c27f-487e-98d8-c840a526d40e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.343641] env[62552]: DEBUG oslo_concurrency.lockutils [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] Lock "505d7f8e-c27f-487e-98d8-c840a526d40e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.343796] env[62552]: DEBUG nova.compute.manager [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] No waiting events found dispatching network-vif-plugged-7734d400-5df3-4f55-b6da-b52dfd148993 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 979.343971] env[62552]: WARNING nova.compute.manager [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Received unexpected event network-vif-plugged-7734d400-5df3-4f55-b6da-b52dfd148993 for instance with vm_state building and task_state spawning. [ 979.345230] env[62552]: DEBUG nova.compute.manager [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Received event network-changed-7734d400-5df3-4f55-b6da-b52dfd148993 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 979.345430] env[62552]: DEBUG nova.compute.manager [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Refreshing instance network info cache due to event network-changed-7734d400-5df3-4f55-b6da-b52dfd148993. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 979.345620] env[62552]: DEBUG oslo_concurrency.lockutils [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] Acquiring lock "refresh_cache-505d7f8e-c27f-487e-98d8-c840a526d40e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.357942] env[62552]: DEBUG oslo_vmware.api [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 979.357942] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5289a5be-c8cd-2d7f-6aec-5a6a658beffc" [ 979.357942] env[62552]: _type = "Task" [ 979.357942] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.368140] env[62552]: DEBUG oslo_vmware.api [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5289a5be-c8cd-2d7f-6aec-5a6a658beffc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.452121] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239837, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.453111] env[62552]: DEBUG nova.network.neutron [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Updating instance_info_cache with network_info: [{"id": "7734d400-5df3-4f55-b6da-b52dfd148993", "address": "fa:16:3e:3a:67:98", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7734d400-5d", "ovs_interfaceid": "7734d400-5df3-4f55-b6da-b52dfd148993", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.631940] env[62552]: DEBUG nova.compute.manager [req-f1177529-8b06-4839-9e03-236a1ef208f7 req-979a3253-a6a0-44ac-ab50-e276f9a75c4a service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received event network-vif-plugged-34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 979.631940] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1177529-8b06-4839-9e03-236a1ef208f7 req-979a3253-a6a0-44ac-ab50-e276f9a75c4a service nova] Acquiring lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.631940] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1177529-8b06-4839-9e03-236a1ef208f7 req-979a3253-a6a0-44ac-ab50-e276f9a75c4a service nova] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.632590] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1177529-8b06-4839-9e03-236a1ef208f7 req-979a3253-a6a0-44ac-ab50-e276f9a75c4a service nova] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.632590] env[62552]: DEBUG nova.compute.manager [req-f1177529-8b06-4839-9e03-236a1ef208f7 req-979a3253-a6a0-44ac-ab50-e276f9a75c4a service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] No waiting events found dispatching network-vif-plugged-34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 979.632590] env[62552]: WARNING nova.compute.manager [req-f1177529-8b06-4839-9e03-236a1ef208f7 req-979a3253-a6a0-44ac-ab50-e276f9a75c4a service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received unexpected event network-vif-plugged-34124ec0-5d30-4c38-8095-b11bda9f58fa for instance with vm_state building and task_state spawning. [ 979.672028] env[62552]: DEBUG oslo_vmware.api [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239838, 'name': PowerOffVM_Task, 'duration_secs': 0.336404} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.672028] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.672028] env[62552]: DEBUG nova.compute.manager [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 979.672028] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06cee08d-881a-428a-a14a-5a42ac36ce0b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.684774] env[62552]: DEBUG oslo_concurrency.lockutils [None req-717ced0d-b708-45b3-b231-93b16d41e6b2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "0418260a-aa27-4955-ab15-b180ec04f0b3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.653s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.756087] env[62552]: DEBUG nova.network.neutron [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Successfully updated port: 34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.760417] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239836, 'name': ReconfigVM_Task, 'duration_secs': 0.864684} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.760953] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734/2c8a5f30-b54f-42ff-a5e0-04eceb1ac734.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.761705] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-05060b67-b419-4691-b0c3-e78cff293110 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.771740] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 979.771740] env[62552]: value = "task-1239839" [ 979.771740] env[62552]: _type = "Task" [ 979.771740] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.782641] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239839, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.867747] env[62552]: DEBUG oslo_vmware.api [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5289a5be-c8cd-2d7f-6aec-5a6a658beffc, 'name': SearchDatastore_Task, 'duration_secs': 0.025852} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.868139] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.950589] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239837, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.900957} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.951583] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 304ad9f2-1965-4a70-aec9-0d2c931b8d95/304ad9f2-1965-4a70-aec9-0d2c931b8d95.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 979.951583] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.951583] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-61873c04-9f92-4aa8-82b2-6eff27065329 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.955943] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-505d7f8e-c27f-487e-98d8-c840a526d40e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.956249] env[62552]: DEBUG nova.compute.manager [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Instance network_info: |[{"id": "7734d400-5df3-4f55-b6da-b52dfd148993", "address": "fa:16:3e:3a:67:98", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7734d400-5d", "ovs_interfaceid": "7734d400-5df3-4f55-b6da-b52dfd148993", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 979.956523] env[62552]: DEBUG oslo_concurrency.lockutils [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] Acquired lock "refresh_cache-505d7f8e-c27f-487e-98d8-c840a526d40e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.956702] env[62552]: DEBUG nova.network.neutron [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Refreshing network info cache for port 7734d400-5df3-4f55-b6da-b52dfd148993 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.958029] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:67:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7734d400-5df3-4f55-b6da-b52dfd148993', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.965268] env[62552]: DEBUG oslo.service.loopingcall [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.969306] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 979.969659] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 979.969659] env[62552]: value = "task-1239840" [ 979.969659] env[62552]: _type = "Task" [ 979.969659] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.970131] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-717eaa12-d7ce-41a7-928c-1f49ebf4f473 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.992372] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239840, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.993503] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.993503] env[62552]: value = "task-1239841" [ 979.993503] env[62552]: _type = "Task" [ 979.993503] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.001052] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239841, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.182372] env[62552]: DEBUG nova.network.neutron [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Updated VIF entry in instance network info cache for port 7734d400-5df3-4f55-b6da-b52dfd148993. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.182736] env[62552]: DEBUG nova.network.neutron [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Updating instance_info_cache with network_info: [{"id": "7734d400-5df3-4f55-b6da-b52dfd148993", "address": "fa:16:3e:3a:67:98", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7734d400-5d", "ovs_interfaceid": "7734d400-5df3-4f55-b6da-b52dfd148993", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.190573] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2b18685d-292c-442a-a43a-da84d2f5723f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.061s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.262247] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.262399] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.262550] env[62552]: DEBUG nova.network.neutron [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 980.282030] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239839, 'name': Rename_Task, 'duration_secs': 0.399} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.282030] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 980.282595] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f293ee9-afd6-44db-b720-25fe22d18b7d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.291291] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 980.291291] env[62552]: value = "task-1239842" [ 980.291291] env[62552]: _type = "Task" [ 980.291291] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.299796] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239842, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.384097] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a015904d-eada-4ac7-a51e-0cfcb4495002 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.392894] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c6853b-5275-421b-80e1-51a6e4c1d404 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.425114] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc94de4-165e-4981-ad26-a9432a0d68fb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.432891] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6abea0f-8d26-4879-93b2-1e62d913257d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.447824] env[62552]: DEBUG nova.compute.provider_tree [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.493258] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239840, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103081} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.493637] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.494882] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d84bec2-01d7-4586-b863-3616a2455402 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.525291] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 304ad9f2-1965-4a70-aec9-0d2c931b8d95/304ad9f2-1965-4a70-aec9-0d2c931b8d95.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.527608] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0422966f-ddcc-49bd-8915-e7a47ec1d944 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.546314] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239841, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.552203] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 980.552203] env[62552]: value = "task-1239843" [ 980.552203] env[62552]: _type = "Task" [ 980.552203] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.562722] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239843, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.686456] env[62552]: DEBUG oslo_concurrency.lockutils [req-d98fd227-223f-4a0e-8a34-80faf0968410 req-2834092a-35bc-49a7-8b2c-0f7601e8dc55 service nova] Releasing lock "refresh_cache-505d7f8e-c27f-487e-98d8-c840a526d40e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.801694] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239842, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.817899] env[62552]: DEBUG nova.network.neutron [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 980.980358] env[62552]: ERROR nova.scheduler.client.report [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [req-460ef242-2e13-457b-9cf2-f6158cbd7707] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-460ef242-2e13-457b-9cf2-f6158cbd7707"}]} [ 980.999906] env[62552]: DEBUG nova.scheduler.client.report [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 981.007178] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239841, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.020779] env[62552]: DEBUG nova.scheduler.client.report [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 981.021203] env[62552]: DEBUG nova.compute.provider_tree [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.034131] env[62552]: DEBUG nova.scheduler.client.report [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 981.063607] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239843, 'name': ReconfigVM_Task, 'duration_secs': 0.401345} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.064813] env[62552]: DEBUG nova.scheduler.client.report [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 981.067508] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 304ad9f2-1965-4a70-aec9-0d2c931b8d95/304ad9f2-1965-4a70-aec9-0d2c931b8d95.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.068907] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8df97fec-0b38-4129-ab6b-374d56c45f4b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.075543] env[62552]: DEBUG nova.network.neutron [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [{"id": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "address": "fa:16:3e:c9:01:94", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34124ec0-5d", "ovs_interfaceid": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.078113] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 981.078113] env[62552]: value = "task-1239844" [ 981.078113] env[62552]: _type = "Task" [ 981.078113] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.088879] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239844, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.120224] env[62552]: INFO nova.compute.manager [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Rebuilding instance [ 981.170288] env[62552]: DEBUG nova.compute.manager [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 981.171162] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fa8e69-f95f-4840-a260-8fa4e1fd6f93 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.301972] env[62552]: DEBUG oslo_vmware.api [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239842, 'name': PowerOnVM_Task, 'duration_secs': 0.56999} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.305275] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 981.305494] env[62552]: INFO nova.compute.manager [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Took 9.28 seconds to spawn the instance on the hypervisor. [ 981.305678] env[62552]: DEBUG nova.compute.manager [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 981.310901] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15917e57-e74f-4749-a809-8d47b38f158c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.329039] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80987e4-1bb3-490a-bedd-6c7e95b523a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.336501] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9162435-394b-479c-99b3-8fa0cd418dc9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.372028] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5fea7c-f768-4f40-8bad-166cc42cf66c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.379276] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-552e193b-190c-441d-8669-30f9d0f57368 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.392851] env[62552]: DEBUG nova.compute.provider_tree [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.504258] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239841, 'name': CreateVM_Task, 'duration_secs': 1.383885} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.504432] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 981.505152] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.505328] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.505673] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 981.505954] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33e474f9-9ecb-4f06-80fa-f33127a18a96 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.510686] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 981.510686] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524e66c0-634f-da12-3bb1-7cb6bbceec92" [ 981.510686] env[62552]: _type = "Task" [ 981.510686] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.518129] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524e66c0-634f-da12-3bb1-7cb6bbceec92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.579378] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.579740] env[62552]: DEBUG nova.compute.manager [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Instance network_info: |[{"id": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "address": "fa:16:3e:c9:01:94", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34124ec0-5d", "ovs_interfaceid": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 981.580311] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:01:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0a3f99df-d1bc-4a37-a048-263445d4a7b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34124ec0-5d30-4c38-8095-b11bda9f58fa', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 981.587946] env[62552]: DEBUG oslo.service.loopingcall [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.591760] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 981.591760] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7105cdb-94b4-4d17-b598-af6081222015 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.610855] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239844, 'name': Rename_Task, 'duration_secs': 0.172802} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.611624] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.611901] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2042f334-f2b8-45f1-a746-fd04eef62a89 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.614298] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.614298] env[62552]: value = "task-1239845" [ 981.614298] env[62552]: _type = "Task" [ 981.614298] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.620030] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 981.620030] env[62552]: value = "task-1239846" [ 981.620030] env[62552]: _type = "Task" [ 981.620030] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.622843] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239845, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.629996] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.658321] env[62552]: DEBUG nova.compute.manager [req-f0599fad-5d67-4386-a3b1-37be32bb43a9 req-0b0c55da-1c8d-4ea8-b0f4-7508a86a66b9 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received event network-changed-34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 981.658321] env[62552]: DEBUG nova.compute.manager [req-f0599fad-5d67-4386-a3b1-37be32bb43a9 req-0b0c55da-1c8d-4ea8-b0f4-7508a86a66b9 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing instance network info cache due to event network-changed-34124ec0-5d30-4c38-8095-b11bda9f58fa. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 981.658321] env[62552]: DEBUG oslo_concurrency.lockutils [req-f0599fad-5d67-4386-a3b1-37be32bb43a9 req-0b0c55da-1c8d-4ea8-b0f4-7508a86a66b9 service nova] Acquiring lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.658321] env[62552]: DEBUG oslo_concurrency.lockutils [req-f0599fad-5d67-4386-a3b1-37be32bb43a9 req-0b0c55da-1c8d-4ea8-b0f4-7508a86a66b9 service nova] Acquired lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.658321] env[62552]: DEBUG nova.network.neutron [req-f0599fad-5d67-4386-a3b1-37be32bb43a9 req-0b0c55da-1c8d-4ea8-b0f4-7508a86a66b9 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing network info cache for port 34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 981.823436] env[62552]: INFO nova.compute.manager [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Took 25.50 seconds to build instance. [ 981.940145] env[62552]: DEBUG nova.scheduler.client.report [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 981.940512] env[62552]: DEBUG nova.compute.provider_tree [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 115 to 116 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 981.940735] env[62552]: DEBUG nova.compute.provider_tree [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 982.023888] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524e66c0-634f-da12-3bb1-7cb6bbceec92, 'name': SearchDatastore_Task, 'duration_secs': 0.011167} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.024235] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.024474] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.024800] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.024944] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.025140] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.025801] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f4f21a2-37d1-43e8-8d77-f16e9fbfa42d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.036536] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.037741] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 982.038686] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d1f0e02-0767-4f08-8200-3ea7cd15d99b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.045135] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 982.045135] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52994a38-4652-271f-2767-53318deb0be9" [ 982.045135] env[62552]: _type = "Task" [ 982.045135] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.052839] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52994a38-4652-271f-2767-53318deb0be9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.125126] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239845, 'name': CreateVM_Task, 'duration_secs': 0.431459} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.128358] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 982.129146] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.129334] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.129663] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.130360] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0406d6a-41d7-45d3-badf-427276389a08 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.135771] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239846, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.138642] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 982.138642] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fe3360-36d1-c49e-0bab-ba4a44151988" [ 982.138642] env[62552]: _type = "Task" [ 982.138642] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.145878] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fe3360-36d1-c49e-0bab-ba4a44151988, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.186986] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.186986] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06595f93-9c75-48ce-9162-5bf6db96e118 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.195948] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 982.195948] env[62552]: value = "task-1239847" [ 982.195948] env[62552]: _type = "Task" [ 982.195948] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.203920] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239847, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.325788] env[62552]: DEBUG oslo_concurrency.lockutils [None req-89226b95-6aa5-4f49-b8ea-df9dafc319e2 tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.016s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.357848] env[62552]: DEBUG nova.network.neutron [req-f0599fad-5d67-4386-a3b1-37be32bb43a9 req-0b0c55da-1c8d-4ea8-b0f4-7508a86a66b9 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updated VIF entry in instance network info cache for port 34124ec0-5d30-4c38-8095-b11bda9f58fa. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 982.358290] env[62552]: DEBUG nova.network.neutron [req-f0599fad-5d67-4386-a3b1-37be32bb43a9 req-0b0c55da-1c8d-4ea8-b0f4-7508a86a66b9 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [{"id": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "address": "fa:16:3e:c9:01:94", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34124ec0-5d", "ovs_interfaceid": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.446860] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.302s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.447398] env[62552]: DEBUG nova.compute.manager [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 982.449931] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.851s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.452136] env[62552]: INFO nova.compute.claims [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.537317] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "38bf968c-e757-4b46-b069-f625f5041c84" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.537538] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.555858] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52994a38-4652-271f-2767-53318deb0be9, 'name': SearchDatastore_Task, 'duration_secs': 0.013539} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.557594] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-009bd88c-c78e-4941-a7b9-faf9bc2a6ac5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.564044] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 982.564044] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5214089f-5bff-ae2c-fd52-8eb9bea5f469" [ 982.564044] env[62552]: _type = "Task" [ 982.564044] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.570354] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5214089f-5bff-ae2c-fd52-8eb9bea5f469, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.633998] env[62552]: DEBUG oslo_vmware.api [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239846, 'name': PowerOnVM_Task, 'duration_secs': 0.714696} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.634302] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 982.634505] env[62552]: INFO nova.compute.manager [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Took 8.20 seconds to spawn the instance on the hypervisor. [ 982.634686] env[62552]: DEBUG nova.compute.manager [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 982.635492] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec028972-b69d-4b3a-a49f-6f70c97f906d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.651849] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fe3360-36d1-c49e-0bab-ba4a44151988, 'name': SearchDatastore_Task, 'duration_secs': 0.012516} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.651955] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.652210] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.652422] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.706915] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 982.707198] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 982.707932] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb076b89-a596-4776-9cb3-2d4d7e6af783 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.714566] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 982.714854] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27d23f36-9bb8-4a43-9b52-672588fd7ddd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.777431] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 982.777675] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 982.777864] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleting the datastore file [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 982.778151] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10eb0076-10d4-46a6-930a-1f908585a8ca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.784777] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 982.784777] env[62552]: value = "task-1239849" [ 982.784777] env[62552]: _type = "Task" [ 982.784777] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.792649] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239849, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.861408] env[62552]: DEBUG oslo_concurrency.lockutils [req-f0599fad-5d67-4386-a3b1-37be32bb43a9 req-0b0c55da-1c8d-4ea8-b0f4-7508a86a66b9 service nova] Releasing lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.956962] env[62552]: DEBUG nova.compute.utils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 982.961905] env[62552]: DEBUG nova.compute.manager [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 982.962198] env[62552]: DEBUG nova.network.neutron [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.040016] env[62552]: DEBUG nova.compute.manager [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 983.076012] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5214089f-5bff-ae2c-fd52-8eb9bea5f469, 'name': SearchDatastore_Task, 'duration_secs': 0.010077} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.076184] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.076464] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 505d7f8e-c27f-487e-98d8-c840a526d40e/505d7f8e-c27f-487e-98d8-c840a526d40e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 983.076764] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.076958] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.077201] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eff50aea-e990-4607-a00f-a0d359586297 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.079393] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49834227-0dbb-4fe4-ae11-e6522f8aaac1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.087036] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 983.087036] env[62552]: value = "task-1239850" [ 983.087036] env[62552]: _type = "Task" [ 983.087036] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.094803] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.094803] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.094803] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dedb1d3-ea7a-41de-8acb-437143942f99 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.099275] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239850, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.102658] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 983.102658] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521414ca-6c34-907a-adcf-fda179f37f09" [ 983.102658] env[62552]: _type = "Task" [ 983.102658] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.111067] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521414ca-6c34-907a-adcf-fda179f37f09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.159367] env[62552]: INFO nova.compute.manager [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Took 21.51 seconds to build instance. [ 983.239818] env[62552]: DEBUG nova.policy [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62819c95c0f14cad89513269b1f0bf1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa13c44376934f519998396c182a49f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.300082] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239849, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13929} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.300496] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.300744] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 983.300994] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 983.462640] env[62552]: DEBUG nova.compute.manager [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 983.535518] env[62552]: DEBUG oslo_concurrency.lockutils [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.535776] env[62552]: DEBUG oslo_concurrency.lockutils [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.564811] env[62552]: DEBUG nova.network.neutron [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Successfully created port: e3f1462f-f27a-413e-b321-ec94895ff702 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 983.569358] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.597884] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239850, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506967} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.603764] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 505d7f8e-c27f-487e-98d8-c840a526d40e/505d7f8e-c27f-487e-98d8-c840a526d40e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 983.603764] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 983.603764] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6771ae82-cc95-47ec-9aa4-4ffb7300a127 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.613093] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521414ca-6c34-907a-adcf-fda179f37f09, 'name': SearchDatastore_Task, 'duration_secs': 0.008059} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.618717] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 983.618717] env[62552]: value = "task-1239851" [ 983.618717] env[62552]: _type = "Task" [ 983.618717] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.618717] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09fbebfc-9702-4a1f-b488-a9a6c39c6af9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.629024] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 983.629024] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52005bd0-5b02-99ab-9c51-c1514496a32c" [ 983.629024] env[62552]: _type = "Task" [ 983.629024] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.632447] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239851, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.644871] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52005bd0-5b02-99ab-9c51-c1514496a32c, 'name': SearchDatastore_Task, 'duration_secs': 0.010933} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.645050] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.645269] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] d0aa144e-9bcf-4faa-8d4c-7743118fbe69/d0aa144e-9bcf-4faa-8d4c-7743118fbe69.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 983.645564] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68e1ec83-e3b7-4fa0-8645-8d5376d1aa8e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.654617] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 983.654617] env[62552]: value = "task-1239852" [ 983.654617] env[62552]: _type = "Task" [ 983.654617] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.662124] env[62552]: DEBUG oslo_concurrency.lockutils [None req-260f2750-7024-419c-8ba4-3e199f386663 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.036s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.669590] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239852, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.790848] env[62552]: DEBUG nova.compute.manager [req-2db4b3ee-e1b3-4f01-b99b-9a3fbf56c1f5 req-c36d0b47-873b-433e-ac40-b59eb333af55 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Received event network-changed-4294380b-fd37-48f3-ba91-6e169312e2a0 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 983.791072] env[62552]: DEBUG nova.compute.manager [req-2db4b3ee-e1b3-4f01-b99b-9a3fbf56c1f5 req-c36d0b47-873b-433e-ac40-b59eb333af55 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Refreshing instance network info cache due to event network-changed-4294380b-fd37-48f3-ba91-6e169312e2a0. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 983.791294] env[62552]: DEBUG oslo_concurrency.lockutils [req-2db4b3ee-e1b3-4f01-b99b-9a3fbf56c1f5 req-c36d0b47-873b-433e-ac40-b59eb333af55 service nova] Acquiring lock "refresh_cache-2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.791443] env[62552]: DEBUG oslo_concurrency.lockutils [req-2db4b3ee-e1b3-4f01-b99b-9a3fbf56c1f5 req-c36d0b47-873b-433e-ac40-b59eb333af55 service nova] Acquired lock "refresh_cache-2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.791608] env[62552]: DEBUG nova.network.neutron [req-2db4b3ee-e1b3-4f01-b99b-9a3fbf56c1f5 req-c36d0b47-873b-433e-ac40-b59eb333af55 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Refreshing network info cache for port 4294380b-fd37-48f3-ba91-6e169312e2a0 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 983.821477] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a639943-f046-4233-86a9-04377388df89 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.836249] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e87bf67-7e5e-4974-a5e4-49507533bab2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.871102] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a362868-83c7-4acb-95fe-9e03434916f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.880705] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9d3f28-c7a7-46bf-bddc-a14ff7375034 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.900021] env[62552]: DEBUG nova.compute.provider_tree [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.039420] env[62552]: INFO nova.compute.manager [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Detaching volume 76baa0b3-5dd2-4828-8f24-0aa4063096c8 [ 984.082168] env[62552]: INFO nova.virt.block_device [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Attempting to driver detach volume 76baa0b3-5dd2-4828-8f24-0aa4063096c8 from mountpoint /dev/sdb [ 984.082461] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 984.082660] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267476', 'volume_id': '76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'name': 'volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0', 'attached_at': '', 'detached_at': '', 'volume_id': '76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'serial': '76baa0b3-5dd2-4828-8f24-0aa4063096c8'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 984.083621] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3394b0-9b34-4927-a7dc-32702f78ec0a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.106085] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cee12a-d416-492d-9aec-ab7422f3446d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.113359] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8ae412-ee67-4bdc-9666-ee4ee76a2fff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.135995] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3811ad-9d0f-4fc5-ba7b-f8687f0b2cbf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.143144] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239851, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090788} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.154733] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 984.155496] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] The volume has not been displaced from its original location: [datastore2] volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8/volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 984.160748] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 984.161581] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185f30cd-2a40-4f02-8be4-986fbb676d5b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.167084] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d9742fc-c6dd-4351-afd1-a0f96f6d7c07 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.204893] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 505d7f8e-c27f-487e-98d8-c840a526d40e/505d7f8e-c27f-487e-98d8-c840a526d40e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.205297] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239852, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532474} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.206785] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29303373-d60a-422a-99c6-1cd4b1708546 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.221669] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] d0aa144e-9bcf-4faa-8d4c-7743118fbe69/d0aa144e-9bcf-4faa-8d4c-7743118fbe69.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 984.221932] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 984.222390] env[62552]: DEBUG oslo_vmware.api [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 984.222390] env[62552]: value = "task-1239853" [ 984.222390] env[62552]: _type = "Task" [ 984.222390] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.222585] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e8696706-f0a6-4743-a534-3834a83b7e21 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.229794] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 984.229794] env[62552]: value = "task-1239854" [ 984.229794] env[62552]: _type = "Task" [ 984.229794] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.233167] env[62552]: DEBUG oslo_vmware.api [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239853, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.237043] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 984.237043] env[62552]: value = "task-1239855" [ 984.237043] env[62552]: _type = "Task" [ 984.237043] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.243472] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239854, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.248199] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239855, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.338047] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.338402] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.338615] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.338869] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.339094] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.339311] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.339612] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.339862] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.340219] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.340219] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.340603] env[62552]: DEBUG nova.virt.hardware [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.341502] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddb5d17-1d97-4757-857e-6989f68592d3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.350857] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9878c0e7-a725-46b4-b14f-8294cf4b872d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.364989] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:f7:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1557e27-81a5-4443-a4f2-15be22c30b3c', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.372373] env[62552]: DEBUG oslo.service.loopingcall [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.374884] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 984.375149] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f2aefc5-cb63-4153-9e11-bf9b1c51fa6b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.394426] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.394426] env[62552]: value = "task-1239856" [ 984.394426] env[62552]: _type = "Task" [ 984.394426] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.405776] env[62552]: DEBUG nova.scheduler.client.report [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 984.409141] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.409373] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.409547] env[62552]: DEBUG nova.compute.manager [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 984.409777] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239856, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.410747] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b89f3f-187f-40cc-807c-e53f2250dff5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.416890] env[62552]: DEBUG nova.compute.manager [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62552) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 984.417464] env[62552]: DEBUG nova.objects.instance [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'flavor' on Instance uuid 304ad9f2-1965-4a70-aec9-0d2c931b8d95 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.479965] env[62552]: DEBUG nova.compute.manager [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 984.503550] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.503868] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.504084] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.504290] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.504456] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.504622] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.504858] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.505056] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.505261] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.505402] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.505578] env[62552]: DEBUG nova.virt.hardware [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.506476] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9fef5b-26ba-4770-ba4b-6dd6a5f77ec5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.518220] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82645ae8-03a0-4d8e-941f-3ca69a9887e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.547011] env[62552]: DEBUG nova.network.neutron [req-2db4b3ee-e1b3-4f01-b99b-9a3fbf56c1f5 req-c36d0b47-873b-433e-ac40-b59eb333af55 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Updated VIF entry in instance network info cache for port 4294380b-fd37-48f3-ba91-6e169312e2a0. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 984.547427] env[62552]: DEBUG nova.network.neutron [req-2db4b3ee-e1b3-4f01-b99b-9a3fbf56c1f5 req-c36d0b47-873b-433e-ac40-b59eb333af55 service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Updating instance_info_cache with network_info: [{"id": "4294380b-fd37-48f3-ba91-6e169312e2a0", "address": "fa:16:3e:b6:87:b0", "network": {"id": "a4d7c75f-bdf3-41a6-9c2f-347202c7bae9", "bridge": "br-int", "label": "tempest-ServersTestJSON-464632350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ecba34d324f42d7af8d1bcf38bbda05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89f807d9-140f-4a6f-8bce-96795f9482ee", "external-id": "nsx-vlan-transportzone-762", "segmentation_id": 762, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4294380b-fd", "ovs_interfaceid": "4294380b-fd37-48f3-ba91-6e169312e2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.735382] env[62552]: DEBUG oslo_vmware.api [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239853, 'name': ReconfigVM_Task, 'duration_secs': 0.2502} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.738419] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 984.743668] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2ccfe6f-cbba-4a8d-9a2c-f57b724bda10 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.763215] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.766980] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239855, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071973} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.767303] env[62552]: DEBUG oslo_vmware.api [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 984.767303] env[62552]: value = "task-1239857" [ 984.767303] env[62552]: _type = "Task" [ 984.767303] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.767537] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 984.768364] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a5cddb-4c7c-4965-a458-49e9023b9bf4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.778746] env[62552]: DEBUG oslo_vmware.api [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239857, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.797585] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] d0aa144e-9bcf-4faa-8d4c-7743118fbe69/d0aa144e-9bcf-4faa-8d4c-7743118fbe69.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.797965] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-960e06fb-a743-4a4d-9104-7c77962aeb01 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.819112] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 984.819112] env[62552]: value = "task-1239858" [ 984.819112] env[62552]: _type = "Task" [ 984.819112] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.829113] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239858, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.904645] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239856, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.911920] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.912530] env[62552]: DEBUG nova.compute.manager [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 984.915593] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.994s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.915868] env[62552]: DEBUG nova.objects.instance [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'resources' on Instance uuid cdee22a9-4327-47af-ab14-8403a06b2802 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.968423] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 984.968877] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.050545] env[62552]: DEBUG oslo_concurrency.lockutils [req-2db4b3ee-e1b3-4f01-b99b-9a3fbf56c1f5 req-c36d0b47-873b-433e-ac40-b59eb333af55 service nova] Releasing lock "refresh_cache-2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.128745] env[62552]: DEBUG nova.network.neutron [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Successfully updated port: e3f1462f-f27a-413e-b321-ec94895ff702 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.245875] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239854, 'name': ReconfigVM_Task, 'duration_secs': 0.695537} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.246261] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 505d7f8e-c27f-487e-98d8-c840a526d40e/505d7f8e-c27f-487e-98d8-c840a526d40e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.246764] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8abdb5e8-dfd4-4e46-a87c-e9ca11ecff13 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.253907] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 985.253907] env[62552]: value = "task-1239859" [ 985.253907] env[62552]: _type = "Task" [ 985.253907] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.261828] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239859, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.276488] env[62552]: DEBUG oslo_vmware.api [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239857, 'name': ReconfigVM_Task, 'duration_secs': 0.17539} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.276877] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267476', 'volume_id': '76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'name': 'volume-76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0', 'attached_at': '', 'detached_at': '', 'volume_id': '76baa0b3-5dd2-4828-8f24-0aa4063096c8', 'serial': '76baa0b3-5dd2-4828-8f24-0aa4063096c8'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 985.330095] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239858, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.404335] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239856, 'name': CreateVM_Task, 'duration_secs': 0.69388} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.404540] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.405182] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.405359] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.405690] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.405950] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8ae0242-debe-4c8d-9f8a-300f1e5c6f42 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.410636] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 985.410636] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52754e0b-04a3-d1ce-c43e-9816eb7c0762" [ 985.410636] env[62552]: _type = "Task" [ 985.410636] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.420354] env[62552]: DEBUG nova.compute.utils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.423816] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52754e0b-04a3-d1ce-c43e-9816eb7c0762, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.424478] env[62552]: DEBUG nova.compute.manager [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 985.424649] env[62552]: DEBUG nova.network.neutron [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.426443] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 985.426602] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36a61a1a-06e9-4634-ba21-374bdb0aa5dc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.432771] env[62552]: DEBUG oslo_vmware.api [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 985.432771] env[62552]: value = "task-1239860" [ 985.432771] env[62552]: _type = "Task" [ 985.432771] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.442631] env[62552]: DEBUG oslo_vmware.api [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239860, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.470146] env[62552]: DEBUG nova.policy [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '850e0e8f6c3e45cda012f934747cc17b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f44eb24ac74742bdbc1ce860fcfde584', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.476338] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.476561] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 985.631442] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "refresh_cache-c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.631663] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired lock "refresh_cache-c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.631888] env[62552]: DEBUG nova.network.neutron [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 985.656017] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7817b12-16fb-448e-b8be-e655d1956918 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.662026] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b340306-2f89-4483-b1a0-b3f806b9124d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.697027] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e39682c-d225-4e20-8245-2174070b50a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.704576] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a484683-dd0b-47cf-91e0-95310fc56426 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.719726] env[62552]: DEBUG nova.compute.provider_tree [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.764613] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239859, 'name': Rename_Task, 'duration_secs': 0.141341} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.764802] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 985.765068] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1264d2b2-09f8-4042-82c3-649ef95411fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.767309] env[62552]: DEBUG nova.network.neutron [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Successfully created port: 13ae1de6-6606-415b-9afd-0bce65b48717 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.773988] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 985.773988] env[62552]: value = "task-1239861" [ 985.773988] env[62552]: _type = "Task" [ 985.773988] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.782482] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239861, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.821764] env[62552]: DEBUG nova.objects.instance [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.831689] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239858, 'name': ReconfigVM_Task, 'duration_secs': 0.609361} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.831960] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Reconfigured VM instance instance-0000005b to attach disk [datastore2] d0aa144e-9bcf-4faa-8d4c-7743118fbe69/d0aa144e-9bcf-4faa-8d4c-7743118fbe69.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.832902] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d7290f6-01aa-4fa8-bc46-4c3d8be8e83d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.838613] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 985.838613] env[62552]: value = "task-1239862" [ 985.838613] env[62552]: _type = "Task" [ 985.838613] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.847302] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239862, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.868252] env[62552]: DEBUG nova.compute.manager [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Received event network-vif-plugged-e3f1462f-f27a-413e-b321-ec94895ff702 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 985.868510] env[62552]: DEBUG oslo_concurrency.lockutils [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] Acquiring lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.868848] env[62552]: DEBUG oslo_concurrency.lockutils [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] Lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.869485] env[62552]: DEBUG oslo_concurrency.lockutils [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] Lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.869852] env[62552]: DEBUG nova.compute.manager [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] No waiting events found dispatching network-vif-plugged-e3f1462f-f27a-413e-b321-ec94895ff702 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 985.870135] env[62552]: WARNING nova.compute.manager [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Received unexpected event network-vif-plugged-e3f1462f-f27a-413e-b321-ec94895ff702 for instance with vm_state building and task_state spawning. [ 985.870688] env[62552]: DEBUG nova.compute.manager [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Received event network-changed-e3f1462f-f27a-413e-b321-ec94895ff702 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 985.870688] env[62552]: DEBUG nova.compute.manager [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Refreshing instance network info cache due to event network-changed-e3f1462f-f27a-413e-b321-ec94895ff702. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 985.870688] env[62552]: DEBUG oslo_concurrency.lockutils [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] Acquiring lock "refresh_cache-c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.921351] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52754e0b-04a3-d1ce-c43e-9816eb7c0762, 'name': SearchDatastore_Task, 'duration_secs': 0.008696} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.921676] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.921865] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.922164] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.922320] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.922507] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.922797] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac12a7de-c9ac-4dbe-b687-0c11de07c628 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.929081] env[62552]: DEBUG nova.compute.manager [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 985.943853] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.944143] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.945287] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-299cd923-8c48-452d-83c0-8550c57640e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.951130] env[62552]: DEBUG oslo_vmware.api [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239860, 'name': PowerOffVM_Task, 'duration_secs': 0.479078} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.951904] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 985.952234] env[62552]: DEBUG nova.compute.manager [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 985.953056] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9584c7f-8497-4847-95bc-bf33e0f31481 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.957132] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 985.957132] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d47310-21d5-ddb5-e4c7-2f7935d2c76a" [ 985.957132] env[62552]: _type = "Task" [ 985.957132] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.968920] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d47310-21d5-ddb5-e4c7-2f7935d2c76a, 'name': SearchDatastore_Task, 'duration_secs': 0.010978} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.969935] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52e530cd-3a3b-47e4-bb1c-1ac80ceed879 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.978054] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 985.978054] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5269b0b0-3258-ae01-726f-90f63353fdf8" [ 985.978054] env[62552]: _type = "Task" [ 985.978054] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.988566] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5269b0b0-3258-ae01-726f-90f63353fdf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.179166] env[62552]: DEBUG nova.network.neutron [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 986.223308] env[62552]: DEBUG nova.scheduler.client.report [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 986.290620] env[62552]: DEBUG oslo_vmware.api [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239861, 'name': PowerOnVM_Task, 'duration_secs': 0.450907} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.290903] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 986.291209] env[62552]: INFO nova.compute.manager [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Took 9.47 seconds to spawn the instance on the hypervisor. [ 986.291438] env[62552]: DEBUG nova.compute.manager [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 986.292316] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de810686-8fe7-40e3-b42e-794b2a12742b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.350656] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239862, 'name': Rename_Task, 'duration_secs': 0.166401} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.350656] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 986.350656] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7224be99-c80b-4f44-9a03-38dffaf20d60 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.356717] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 986.356717] env[62552]: value = "task-1239863" [ 986.356717] env[62552]: _type = "Task" [ 986.356717] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.365035] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.458449] env[62552]: DEBUG nova.network.neutron [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Updating instance_info_cache with network_info: [{"id": "e3f1462f-f27a-413e-b321-ec94895ff702", "address": "fa:16:3e:2c:4b:50", "network": {"id": "bfa99bf2-fb08-4fb9-9231-78e1258e14a0", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-499016683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa13c44376934f519998396c182a49f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3f1462f-f2", "ovs_interfaceid": "e3f1462f-f27a-413e-b321-ec94895ff702", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.475095] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7c4964c3-fed4-4862-bdb2-523c0f200af1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.065s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.491599] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5269b0b0-3258-ae01-726f-90f63353fdf8, 'name': SearchDatastore_Task, 'duration_secs': 0.010408} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.492208] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.492639] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71/ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 986.493148] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6741835-236f-4234-bafa-1ac4e8998089 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.501410] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 986.501410] env[62552]: value = "task-1239864" [ 986.501410] env[62552]: _type = "Task" [ 986.501410] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.509976] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239864, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.729580] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.814s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.732070] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.733573] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 6.865s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.754739] env[62552]: INFO nova.scheduler.client.report [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted allocations for instance cdee22a9-4327-47af-ab14-8403a06b2802 [ 986.808908] env[62552]: INFO nova.compute.manager [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Took 23.86 seconds to build instance. [ 986.831172] env[62552]: DEBUG oslo_concurrency.lockutils [None req-49c31f46-4958-4091-bc09-b025cd0dbf11 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.295s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.832326] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.100s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.832520] env[62552]: DEBUG nova.compute.manager [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 986.833610] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f197bb03-89a7-4d4c-b16b-94c62350984f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.843051] env[62552]: DEBUG nova.compute.manager [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62552) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 986.843667] env[62552]: DEBUG nova.objects.instance [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.872305] env[62552]: DEBUG oslo_vmware.api [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239863, 'name': PowerOnVM_Task, 'duration_secs': 0.464258} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.872809] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 986.873212] env[62552]: INFO nova.compute.manager [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Took 7.61 seconds to spawn the instance on the hypervisor. [ 986.873788] env[62552]: DEBUG nova.compute.manager [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 986.874829] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd96ffdb-d602-4330-b71d-01fb22572cad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.884744] env[62552]: WARNING oslo_messaging._drivers.amqpdriver [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 986.942951] env[62552]: DEBUG nova.compute.manager [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 986.960477] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Releasing lock "refresh_cache-c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.960850] env[62552]: DEBUG nova.compute.manager [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Instance network_info: |[{"id": "e3f1462f-f27a-413e-b321-ec94895ff702", "address": "fa:16:3e:2c:4b:50", "network": {"id": "bfa99bf2-fb08-4fb9-9231-78e1258e14a0", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-499016683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa13c44376934f519998396c182a49f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3f1462f-f2", "ovs_interfaceid": "e3f1462f-f27a-413e-b321-ec94895ff702", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 986.961230] env[62552]: DEBUG oslo_concurrency.lockutils [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] Acquired lock "refresh_cache-c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.961800] env[62552]: DEBUG nova.network.neutron [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Refreshing network info cache for port e3f1462f-f27a-413e-b321-ec94895ff702 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 986.963720] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:4b:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e3f1462f-f27a-413e-b321-ec94895ff702', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.972331] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Creating folder: Project (aa13c44376934f519998396c182a49f2). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 986.975553] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9a5811f-7722-4b19-a328-6cc0e3166327 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.985088] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.985383] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.985578] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.985794] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.985935] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.986301] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.986472] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.986641] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.986814] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.986991] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.987224] env[62552]: DEBUG nova.virt.hardware [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.990367] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af51a64d-5f2d-4c1f-9606-76ecefa2df27 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.994898] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Created folder: Project (aa13c44376934f519998396c182a49f2) in parent group-v267339. [ 986.995111] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Creating folder: Instances. Parent ref: group-v267492. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 986.995727] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c4a7423-2c53-481a-bbb8-980c764af25c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.001280] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f38be70-4005-48d2-b6ea-5701ad5dc207 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.010222] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Created folder: Instances in parent group-v267492. [ 987.010502] env[62552]: DEBUG oslo.service.loopingcall [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.019228] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.020374] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da5b569a-ad0a-4987-8423-7495529e2099 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.039958] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239864, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.046265] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.046265] env[62552]: value = "task-1239867" [ 987.046265] env[62552]: _type = "Task" [ 987.046265] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.053802] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239867, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.142698] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.142975] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.143199] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.143444] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.143560] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.146224] env[62552]: INFO nova.compute.manager [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Terminating instance [ 987.202350] env[62552]: DEBUG nova.compute.manager [req-0fdf6775-74d1-4d74-976f-4ffdb0646baf req-09380334-5366-4cfe-96e0-a34d674e9fb4 service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Received event network-vif-plugged-13ae1de6-6606-415b-9afd-0bce65b48717 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 987.202570] env[62552]: DEBUG oslo_concurrency.lockutils [req-0fdf6775-74d1-4d74-976f-4ffdb0646baf req-09380334-5366-4cfe-96e0-a34d674e9fb4 service nova] Acquiring lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.202878] env[62552]: DEBUG oslo_concurrency.lockutils [req-0fdf6775-74d1-4d74-976f-4ffdb0646baf req-09380334-5366-4cfe-96e0-a34d674e9fb4 service nova] Lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.203073] env[62552]: DEBUG oslo_concurrency.lockutils [req-0fdf6775-74d1-4d74-976f-4ffdb0646baf req-09380334-5366-4cfe-96e0-a34d674e9fb4 service nova] Lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.203257] env[62552]: DEBUG nova.compute.manager [req-0fdf6775-74d1-4d74-976f-4ffdb0646baf req-09380334-5366-4cfe-96e0-a34d674e9fb4 service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] No waiting events found dispatching network-vif-plugged-13ae1de6-6606-415b-9afd-0bce65b48717 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 987.203430] env[62552]: WARNING nova.compute.manager [req-0fdf6775-74d1-4d74-976f-4ffdb0646baf req-09380334-5366-4cfe-96e0-a34d674e9fb4 service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Received unexpected event network-vif-plugged-13ae1de6-6606-415b-9afd-0bce65b48717 for instance with vm_state building and task_state spawning. [ 987.264328] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f7d826f2-5b0e-4b1c-aee0-55c69b585e1b tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "cdee22a9-4327-47af-ab14-8403a06b2802" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.632s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.310916] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ec140c11-b9df-4098-85c9-59e0bdf23355 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "505d7f8e-c27f-487e-98d8-c840a526d40e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.370s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.400665] env[62552]: INFO nova.compute.manager [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Took 22.01 seconds to build instance. [ 987.517029] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239864, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.594112} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.518181] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71/ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 987.518413] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.519378] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a28d85-05a2-4d21-8fa9-2040f30769b6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.522350] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00b1c3ff-be61-4e72-9fd6-e285deb4e40d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.530033] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3354f894-6f07-4007-85e7-130763d310ec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.533973] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 987.533973] env[62552]: value = "task-1239868" [ 987.533973] env[62552]: _type = "Task" [ 987.533973] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.575217] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c186adcb-6c07-4ef4-a600-5df3c0579176 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.583537] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239868, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.590482] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3106b6-7ac4-4f19-b4fc-f556f4710483 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.594565] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239867, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.605858] env[62552]: DEBUG nova.compute.provider_tree [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.649972] env[62552]: DEBUG nova.compute.manager [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 987.650243] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 987.651100] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35aada89-859f-46d0-93e1-9bb1dad25f72 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.659287] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 987.659593] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9703078-7c0f-49ff-b9f5-2aa053130d4f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.739549] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 987.740097] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 987.740097] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleting the datastore file [datastore2] 304ad9f2-1965-4a70-aec9-0d2c931b8d95 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.740312] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84f4b288-8f56-4bca-916a-43b6b9067441 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.747813] env[62552]: DEBUG oslo_vmware.api [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 987.747813] env[62552]: value = "task-1239870" [ 987.747813] env[62552]: _type = "Task" [ 987.747813] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.757792] env[62552]: DEBUG oslo_vmware.api [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239870, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.773433] env[62552]: DEBUG nova.network.neutron [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Successfully updated port: 13ae1de6-6606-415b-9afd-0bce65b48717 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.850773] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 987.851096] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4ed1a57-1420-43d3-8f3c-c36bb7707f16 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.858484] env[62552]: DEBUG oslo_vmware.api [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 987.858484] env[62552]: value = "task-1239871" [ 987.858484] env[62552]: _type = "Task" [ 987.858484] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.869152] env[62552]: DEBUG oslo_vmware.api [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239871, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.903730] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c430be17-b03a-491f-b245-ac3e811b6b18 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.522s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.926962] env[62552]: INFO nova.compute.manager [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Rebuilding instance [ 987.977993] env[62552]: DEBUG nova.compute.manager [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 987.978927] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5abb39b-16e1-4c87-9228-3e0f71d98417 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.045015] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239868, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077645} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.045375] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.046214] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d598e6-e105-417a-9511-1f3606d86beb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.069798] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71/ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.070192] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90cabe6d-50a6-4382-bb8f-586a15344bd4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.094399] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239867, 'name': CreateVM_Task, 'duration_secs': 0.599485} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.095740] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.096156] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 988.096156] env[62552]: value = "task-1239872" [ 988.096156] env[62552]: _type = "Task" [ 988.096156] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.096963] env[62552]: DEBUG nova.network.neutron [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Updated VIF entry in instance network info cache for port e3f1462f-f27a-413e-b321-ec94895ff702. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 988.097305] env[62552]: DEBUG nova.network.neutron [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Updating instance_info_cache with network_info: [{"id": "e3f1462f-f27a-413e-b321-ec94895ff702", "address": "fa:16:3e:2c:4b:50", "network": {"id": "bfa99bf2-fb08-4fb9-9231-78e1258e14a0", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-499016683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa13c44376934f519998396c182a49f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3f1462f-f2", "ovs_interfaceid": "e3f1462f-f27a-413e-b321-ec94895ff702", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.102057] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.102057] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.102057] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.102057] env[62552]: DEBUG nova.compute.manager [req-22ad64a7-3924-4748-aa5a-ff3ab21c073c req-cdf16bda-7a23-40e9-9ef3-12e6cb5baffd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Received event network-changed-13ae1de6-6606-415b-9afd-0bce65b48717 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 988.102057] env[62552]: DEBUG nova.compute.manager [req-22ad64a7-3924-4748-aa5a-ff3ab21c073c req-cdf16bda-7a23-40e9-9ef3-12e6cb5baffd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Refreshing instance network info cache due to event network-changed-13ae1de6-6606-415b-9afd-0bce65b48717. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 988.102633] env[62552]: DEBUG oslo_concurrency.lockutils [req-22ad64a7-3924-4748-aa5a-ff3ab21c073c req-cdf16bda-7a23-40e9-9ef3-12e6cb5baffd service nova] Acquiring lock "refresh_cache-0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.102633] env[62552]: DEBUG oslo_concurrency.lockutils [req-22ad64a7-3924-4748-aa5a-ff3ab21c073c req-cdf16bda-7a23-40e9-9ef3-12e6cb5baffd service nova] Acquired lock "refresh_cache-0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.102633] env[62552]: DEBUG nova.network.neutron [req-22ad64a7-3924-4748-aa5a-ff3ab21c073c req-cdf16bda-7a23-40e9-9ef3-12e6cb5baffd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Refreshing network info cache for port 13ae1de6-6606-415b-9afd-0bce65b48717 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.103966] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c09a1079-1222-4e81-8a8f-f478b66801ed {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.113837] env[62552]: DEBUG nova.scheduler.client.report [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 988.121464] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 988.121464] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b6300c-0c5b-0a37-3a8b-bd82ba3d84ff" [ 988.121464] env[62552]: _type = "Task" [ 988.121464] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.126659] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.135924] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b6300c-0c5b-0a37-3a8b-bd82ba3d84ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.196456] env[62552]: DEBUG nova.network.neutron [req-22ad64a7-3924-4748-aa5a-ff3ab21c073c req-cdf16bda-7a23-40e9-9ef3-12e6cb5baffd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 988.258177] env[62552]: DEBUG oslo_vmware.api [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239870, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.389784} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.258445] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.258702] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 988.260421] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 988.260421] env[62552]: INFO nova.compute.manager [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Took 0.61 seconds to destroy the instance on the hypervisor. [ 988.260421] env[62552]: DEBUG oslo.service.loopingcall [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.260421] env[62552]: DEBUG nova.compute.manager [-] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 988.260421] env[62552]: DEBUG nova.network.neutron [-] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 988.276435] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquiring lock "refresh_cache-0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.365672] env[62552]: DEBUG nova.network.neutron [req-22ad64a7-3924-4748-aa5a-ff3ab21c073c req-cdf16bda-7a23-40e9-9ef3-12e6cb5baffd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.370025] env[62552]: DEBUG oslo_vmware.api [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239871, 'name': PowerOffVM_Task, 'duration_secs': 0.387436} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.370802] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.370802] env[62552]: DEBUG nova.compute.manager [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 988.371456] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac5ec11-9b23-450c-b4bd-8ab23eab0050 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.604885] env[62552]: DEBUG oslo_concurrency.lockutils [req-299e2307-4de6-4638-a44e-597aed0362ca req-d151b66f-041a-43d5-b246-e69e69f74710 service nova] Releasing lock "refresh_cache-c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.613780] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.638462] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b6300c-0c5b-0a37-3a8b-bd82ba3d84ff, 'name': SearchDatastore_Task, 'duration_secs': 0.026897} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.638985] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.639475] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.639475] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.639628] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.639781] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.640340] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1877376-4e3a-45cb-84d6-6d34b0bea11a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.649330] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.649486] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.650306] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75530f7f-3deb-4165-b34e-9c88945473ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.655962] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 988.655962] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f52df2-1964-5cdd-2be0-52b8cb1b889a" [ 988.655962] env[62552]: _type = "Task" [ 988.655962] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.664761] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f52df2-1964-5cdd-2be0-52b8cb1b889a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.871635] env[62552]: DEBUG oslo_concurrency.lockutils [req-22ad64a7-3924-4748-aa5a-ff3ab21c073c req-cdf16bda-7a23-40e9-9ef3-12e6cb5baffd service nova] Releasing lock "refresh_cache-0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.872093] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquired lock "refresh_cache-0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.872271] env[62552]: DEBUG nova.network.neutron [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.886752] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4ab325a0-ab91-478c-8ab9-f880241dcb3d tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.054s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.994681] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.994681] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa1cbb77-04f4-43b0-aca6-2d0d00f30e1d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.007332] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 989.007332] env[62552]: value = "task-1239873" [ 989.007332] env[62552]: _type = "Task" [ 989.007332] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.011064] env[62552]: DEBUG nova.network.neutron [-] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.018955] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239873, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.063880] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.064026] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquired lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.064176] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Forcefully refreshing network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 989.110821] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239872, 'name': ReconfigVM_Task, 'duration_secs': 0.860454} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.111129] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Reconfigured VM instance instance-00000057 to attach disk [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71/ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.111793] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5412e03f-27b4-4276-850d-27d1daa9e327 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.118324] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 989.118324] env[62552]: value = "task-1239874" [ 989.118324] env[62552]: _type = "Task" [ 989.118324] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.129154] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.396s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.133014] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239874, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.133188] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.564s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.139075] env[62552]: INFO nova.compute.claims [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 989.170684] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f52df2-1964-5cdd-2be0-52b8cb1b889a, 'name': SearchDatastore_Task, 'duration_secs': 0.03821} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.171603] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e0d354c-fc61-4e67-ad0e-64b9a0f26342 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.180876] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 989.180876] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5242d342-2682-25cb-e7e8-ba6eaf224583" [ 989.180876] env[62552]: _type = "Task" [ 989.180876] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.190091] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5242d342-2682-25cb-e7e8-ba6eaf224583, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.276922] env[62552]: DEBUG nova.compute.manager [req-dff1fc75-0bd6-4b4b-bfec-20e319115dbe req-83ce4fc0-2091-4364-816a-108a527886aa service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 989.277185] env[62552]: DEBUG nova.compute.manager [req-dff1fc75-0bd6-4b4b-bfec-20e319115dbe req-83ce4fc0-2091-4364-816a-108a527886aa service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing instance network info cache due to event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 989.277917] env[62552]: DEBUG oslo_concurrency.lockutils [req-dff1fc75-0bd6-4b4b-bfec-20e319115dbe req-83ce4fc0-2091-4364-816a-108a527886aa service nova] Acquiring lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.277917] env[62552]: DEBUG oslo_concurrency.lockutils [req-dff1fc75-0bd6-4b4b-bfec-20e319115dbe req-83ce4fc0-2091-4364-816a-108a527886aa service nova] Acquired lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.277917] env[62552]: DEBUG nova.network.neutron [req-dff1fc75-0bd6-4b4b-bfec-20e319115dbe req-83ce4fc0-2091-4364-816a-108a527886aa service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.403743] env[62552]: DEBUG nova.network.neutron [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.513877] env[62552]: INFO nova.compute.manager [-] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Took 1.25 seconds to deallocate network for instance. [ 989.526056] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239873, 'name': PowerOffVM_Task, 'duration_secs': 0.187176} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.526567] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 989.526801] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 989.527719] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccf85e7-619d-485e-bf37-d184803303c4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.534873] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 989.535139] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1df1f8a6-6fdf-4f09-8dbc-a2931af9d6fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.554724] env[62552]: DEBUG nova.network.neutron [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Updating instance_info_cache with network_info: [{"id": "13ae1de6-6606-415b-9afd-0bce65b48717", "address": "fa:16:3e:b8:4a:1f", "network": {"id": "5b1bbdf9-0cfe-49da-bcf5-4f8b41744f19", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-106952979-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f44eb24ac74742bdbc1ce860fcfde584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13ae1de6-66", "ovs_interfaceid": "13ae1de6-6606-415b-9afd-0bce65b48717", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.607697] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 989.607934] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 989.608146] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleting the datastore file [datastore2] 505d7f8e-c27f-487e-98d8-c840a526d40e {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.608667] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c531199-fb19-49c7-b526-c48b5036f0a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.615127] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 989.615127] env[62552]: value = "task-1239876" [ 989.615127] env[62552]: _type = "Task" [ 989.615127] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.628095] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.631633] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239874, 'name': Rename_Task, 'duration_secs': 0.211235} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.631904] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.632211] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-019417c8-f986-41da-a8a1-92b75527ba58 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.637841] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 989.637841] env[62552]: value = "task-1239877" [ 989.637841] env[62552]: _type = "Task" [ 989.637841] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.651191] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239877, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.692998] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5242d342-2682-25cb-e7e8-ba6eaf224583, 'name': SearchDatastore_Task, 'duration_secs': 0.020076} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.693441] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.693795] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e/c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 989.694369] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8e9006a-5703-4132-b004-9302004b837c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.704903] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 989.704903] env[62552]: value = "task-1239878" [ 989.704903] env[62552]: _type = "Task" [ 989.704903] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.713624] env[62552]: INFO nova.scheduler.client.report [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted allocation for migration 49939264-6417-4e73-9ac7-4f05edf037aa [ 989.723195] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239878, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.807409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "158d4227-90ca-41ae-821b-efd353928cb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.807409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.934041] env[62552]: DEBUG nova.objects.instance [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.029629] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.050750] env[62552]: DEBUG nova.network.neutron [req-dff1fc75-0bd6-4b4b-bfec-20e319115dbe req-83ce4fc0-2091-4364-816a-108a527886aa service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updated VIF entry in instance network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.051264] env[62552]: DEBUG nova.network.neutron [req-dff1fc75-0bd6-4b4b-bfec-20e319115dbe req-83ce4fc0-2091-4364-816a-108a527886aa service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.059102] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Releasing lock "refresh_cache-0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.059472] env[62552]: DEBUG nova.compute.manager [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Instance network_info: |[{"id": "13ae1de6-6606-415b-9afd-0bce65b48717", "address": "fa:16:3e:b8:4a:1f", "network": {"id": "5b1bbdf9-0cfe-49da-bcf5-4f8b41744f19", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-106952979-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f44eb24ac74742bdbc1ce860fcfde584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13ae1de6-66", "ovs_interfaceid": "13ae1de6-6606-415b-9afd-0bce65b48717", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 990.059950] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:4a:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cde23701-02ca-4cb4-b5a6-d321f8ac9660', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13ae1de6-6606-415b-9afd-0bce65b48717', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.073455] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Creating folder: Project (f44eb24ac74742bdbc1ce860fcfde584). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 990.075842] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a291364-7354-4a28-8c5a-62f8f0b6fa11 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.090442] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Created folder: Project (f44eb24ac74742bdbc1ce860fcfde584) in parent group-v267339. [ 990.090674] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Creating folder: Instances. Parent ref: group-v267495. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 990.090887] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2b05878-4ab5-4494-b584-ef995b892ce5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.106249] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Created folder: Instances in parent group-v267495. [ 990.106249] env[62552]: DEBUG oslo.service.loopingcall [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.106249] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 990.106249] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-beca6773-36be-44bc-92ac-4213499db8de {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.140255] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17581} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.140255] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.140255] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.140255] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.142307] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.142307] env[62552]: value = "task-1239881" [ 990.142307] env[62552]: _type = "Task" [ 990.142307] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.160305] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239877, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.165107] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239881, 'name': CreateVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.219398] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239878, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484927} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.219564] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e/c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 990.219819] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 990.221374] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3ac076c6-8afe-4d85-9e5d-f353a9e6ecd8 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.966s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.222811] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8fbfcfb-9a57-452a-ab47-bbc041ec0666 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.235316] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 990.235316] env[62552]: value = "task-1239882" [ 990.235316] env[62552]: _type = "Task" [ 990.235316] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.247499] env[62552]: DEBUG nova.compute.manager [req-5afbf7fd-3d1d-40d7-8232-ad9bccbb1c29 req-899c3510-e29f-4f53-a068-2adf3493cedd service nova] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Received event network-vif-deleted-54ef4be6-f8fc-458e-a1e0-96a8529e448e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 990.251558] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239882, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.308840] env[62552]: DEBUG nova.compute.manager [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 990.365238] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.436639] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6979da3c-80db-484f-878c-61a3fc00b1ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.445334] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.445514] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.445692] env[62552]: DEBUG nova.network.neutron [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.445871] env[62552]: DEBUG nova.objects.instance [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'info_cache' on Instance uuid a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.447774] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b960baf1-f6e2-4aea-9528-8920e1910013 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.479333] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50094039-b4cd-4edf-9ca8-0c0e6d11c9ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.487710] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a575c37-3a7c-49c1-87b0-743b1058c11d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.502165] env[62552]: DEBUG nova.compute.provider_tree [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.558290] env[62552]: DEBUG oslo_concurrency.lockutils [req-dff1fc75-0bd6-4b4b-bfec-20e319115dbe req-83ce4fc0-2091-4364-816a-108a527886aa service nova] Releasing lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.659924] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239881, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.662777] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239877, 'name': PowerOnVM_Task, 'duration_secs': 0.693594} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.663238] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.663238] env[62552]: DEBUG nova.compute.manager [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 990.663906] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d3aecb-3967-4a5f-a256-e43af81dfaf0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.708308] env[62552]: DEBUG oslo_concurrency.lockutils [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "04186fde-a9ed-415a-bd40-312f0347fcc7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.708588] env[62552]: DEBUG oslo_concurrency.lockutils [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.708842] env[62552]: DEBUG oslo_concurrency.lockutils [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "04186fde-a9ed-415a-bd40-312f0347fcc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.709128] env[62552]: DEBUG oslo_concurrency.lockutils [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.709251] env[62552]: DEBUG oslo_concurrency.lockutils [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.712934] env[62552]: INFO nova.compute.manager [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Terminating instance [ 990.746678] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239882, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099958} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.747228] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.748037] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be28ab70-0e2e-4c37-bad6-6e7f5841acc0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.772857] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e/c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.773512] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44a07a9e-7389-4367-b636-86c21c8fcae2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.492242] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Releasing lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.492626] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updated the network info_cache for instance {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 991.493626] env[62552]: DEBUG nova.objects.base [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 991.495040] env[62552]: DEBUG nova.scheduler.client.report [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 991.500657] env[62552]: INFO nova.compute.manager [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] bringing vm to original state: 'stopped' [ 991.503131] env[62552]: DEBUG nova.compute.manager [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 991.503326] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 991.504688] env[62552]: DEBUG nova.compute.manager [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received event network-changed-34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 991.504871] env[62552]: DEBUG nova.compute.manager [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing instance network info cache due to event network-changed-34124ec0-5d30-4c38-8095-b11bda9f58fa. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 991.505094] env[62552]: DEBUG oslo_concurrency.lockutils [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] Acquiring lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.505244] env[62552]: DEBUG oslo_concurrency.lockutils [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] Acquired lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.505405] env[62552]: DEBUG nova.network.neutron [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing network info cache for port 34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.513036] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.513472] env[62552]: DEBUG nova.compute.manager [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 991.519021] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.519989] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca50fcf-fa7a-4ef3-b665-2b4f43c795ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.522944] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "16785811-1927-408b-9ead-5848a5ffdf24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.523200] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "16785811-1927-408b-9ead-5848a5ffdf24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.524590] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.495s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.524808] env[62552]: DEBUG nova.objects.instance [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'resources' on Instance uuid 304ad9f2-1965-4a70-aec9-0d2c931b8d95 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.526350] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.530282] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.534043] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 991.534043] env[62552]: value = "task-1239883" [ 991.534043] env[62552]: _type = "Task" [ 991.534043] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.534043] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.534043] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.541307] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.541461] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 991.542538] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239881, 'name': CreateVM_Task, 'duration_secs': 0.63929} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.543059] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.545184] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.545409] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.545564] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.545746] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.545887] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.546041] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.547143] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.547143] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.547143] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.547143] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.547143] env[62552]: DEBUG nova.virt.hardware [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.547775] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.547996] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.549402] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 991.549628] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3253034-b2fe-4ea6-b01e-322cea4b3315 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.552343] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be45a8f1-43d0-413e-921f-67733333bfa8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.559333] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.559497] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.559888] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.560226] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.561168] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a79779a6-f868-42e0-b5d1-23c38143c7ec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.570672] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5412ee7e-9e6a-4be2-a751-86e4895871cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.575016] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 991.575016] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52244392-5dd9-8b84-094e-7ad4f8d27fe5" [ 991.575016] env[62552]: _type = "Task" [ 991.575016] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.575302] env[62552]: DEBUG oslo_vmware.api [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 991.575302] env[62552]: value = "task-1239884" [ 991.575302] env[62552]: _type = "Task" [ 991.575302] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.589398] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:67:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7734d400-5df3-4f55-b6da-b52dfd148993', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.596774] env[62552]: DEBUG oslo.service.loopingcall [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.598105] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.598344] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6df5d4a7-93d8-4f28-a5e6-9cacb9fec663 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.618846] env[62552]: DEBUG oslo_vmware.api [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239884, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.619128] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52244392-5dd9-8b84-094e-7ad4f8d27fe5, 'name': SearchDatastore_Task, 'duration_secs': 0.011516} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.619815] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.620079] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.620320] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.620470] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.620652] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.620915] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a67aa2e9-6b20-4678-b35b-01155ce6e9af {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.623855] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.623855] env[62552]: value = "task-1239885" [ 991.623855] env[62552]: _type = "Task" [ 991.623855] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.628764] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.628943] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 991.632125] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-888630ea-9212-4e16-b66a-958c02ef7aac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.634155] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239885, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.636884] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 991.636884] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ab3f36-ee2c-fe68-0471-cc23a3b2deaf" [ 991.636884] env[62552]: _type = "Task" [ 991.636884] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.645771] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ab3f36-ee2c-fe68-0471-cc23a3b2deaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.032254] env[62552]: DEBUG nova.compute.manager [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 992.032710] env[62552]: DEBUG nova.compute.utils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 992.042102] env[62552]: DEBUG nova.compute.manager [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 992.042427] env[62552]: DEBUG nova.network.neutron [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 992.056564] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239883, 'name': ReconfigVM_Task, 'duration_secs': 0.49791} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.057403] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Reconfigured VM instance instance-0000005c to attach disk [datastore2] c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e/c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.059075] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93869842-e86d-47bd-baf4-c297e4ced79c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.061651] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.069845] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 992.069845] env[62552]: value = "task-1239886" [ 992.069845] env[62552]: _type = "Task" [ 992.069845] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.082905] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239886, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.088686] env[62552]: DEBUG oslo_vmware.api [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239884, 'name': PowerOffVM_Task, 'duration_secs': 0.343614} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.091596] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.091596] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 992.092677] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9e27272-b1c6-4746-bca3-1722b1b37a99 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.110692] env[62552]: DEBUG nova.policy [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecfe0d4d977540d99709760db0564cf9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3758acde6e3e4320bf63d4cd1667bfc2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 992.136799] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239885, 'name': CreateVM_Task, 'duration_secs': 0.352745} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.138852] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 992.139605] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.139791] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.140165] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 992.145473] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a528106-917d-40c2-834c-e96c395a972d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.151118] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 992.151118] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5280cdf7-f51f-b959-de66-1bbbade44393" [ 992.151118] env[62552]: _type = "Task" [ 992.151118] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.156284] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ab3f36-ee2c-fe68-0471-cc23a3b2deaf, 'name': SearchDatastore_Task, 'duration_secs': 0.008675} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.163019] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf36eb1f-a2cb-41d4-8e43-88d353586399 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.169631] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 992.169882] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 992.170137] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleting the datastore file [datastore2] 04186fde-a9ed-415a-bd40-312f0347fcc7 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.171018] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ddb13f8c-f8a2-4b96-9ca8-ccb1d533a5f4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.177307] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5280cdf7-f51f-b959-de66-1bbbade44393, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.177631] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 992.177631] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a52b76-6143-7e6e-2703-dcbe8e511357" [ 992.177631] env[62552]: _type = "Task" [ 992.177631] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.185534] env[62552]: DEBUG oslo_vmware.api [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 992.185534] env[62552]: value = "task-1239888" [ 992.185534] env[62552]: _type = "Task" [ 992.185534] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.192167] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a52b76-6143-7e6e-2703-dcbe8e511357, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.199465] env[62552]: DEBUG oslo_vmware.api [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239888, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.271988] env[62552]: DEBUG nova.network.neutron [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updating instance_info_cache with network_info: [{"id": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "address": "fa:16:3e:fc:98:ec", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56e2a627-19", "ovs_interfaceid": "56e2a627-1966-4c66-b733-d9e3cf3435ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.335383] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33ade18-b574-43fa-b301-7d7711110e78 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.343799] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780fb26e-2883-4180-98a6-d13e72904382 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.381635] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e33efa7-8ba9-4590-bb59-87b9e132cf85 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.388017] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129d7abf-558d-4088-8fc5-5543529ecb4e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.402442] env[62552]: DEBUG nova.compute.provider_tree [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.516231] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.516602] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.516706] env[62552]: DEBUG nova.compute.manager [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 992.517919] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b906d4e2-07d2-412e-86ae-b6d183129c61 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.526630] env[62552]: DEBUG nova.compute.manager [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62552) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 992.537945] env[62552]: DEBUG nova.compute.manager [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 992.553403] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.580220] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239886, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.628501] env[62552]: DEBUG nova.network.neutron [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updated VIF entry in instance network info cache for port 34124ec0-5d30-4c38-8095-b11bda9f58fa. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.628888] env[62552]: DEBUG nova.network.neutron [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [{"id": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "address": "fa:16:3e:c9:01:94", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34124ec0-5d", "ovs_interfaceid": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.665068] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5280cdf7-f51f-b959-de66-1bbbade44393, 'name': SearchDatastore_Task, 'duration_secs': 0.025053} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.666039] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.666039] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.666039] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.666039] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.666308] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.666466] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21ecad3a-8bc3-4bfd-938b-2e2ccefaeb22 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.676487] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.676693] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 992.677479] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1eacb1b-111e-4942-bc50-ad65482aa4c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.693064] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a52b76-6143-7e6e-2703-dcbe8e511357, 'name': SearchDatastore_Task, 'duration_secs': 0.018132} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.693475] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 992.693475] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529137b0-a08b-056e-8b96-5e96753a9b15" [ 992.693475] env[62552]: _type = "Task" [ 992.693475] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.694206] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.694533] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9/0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 992.694890] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9305ba38-3cc5-49e4-82ee-ae8312df8e81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.703014] env[62552]: DEBUG oslo_vmware.api [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239888, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.512508} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.703893] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.704126] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.704326] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 992.704510] env[62552]: INFO nova.compute.manager [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Took 1.20 seconds to destroy the instance on the hypervisor. [ 992.704752] env[62552]: DEBUG oslo.service.loopingcall [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.705300] env[62552]: DEBUG nova.compute.manager [-] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 992.705427] env[62552]: DEBUG nova.network.neutron [-] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.710334] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529137b0-a08b-056e-8b96-5e96753a9b15, 'name': SearchDatastore_Task, 'duration_secs': 0.011175} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.712411] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 992.712411] env[62552]: value = "task-1239889" [ 992.712411] env[62552]: _type = "Task" [ 992.712411] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.712611] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76f9111a-d6b6-4716-8a4e-14ef04047a70 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.722723] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 992.722723] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528ac194-2831-0b09-879e-68bb618d6d6d" [ 992.722723] env[62552]: _type = "Task" [ 992.722723] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.725788] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239889, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.734091] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528ac194-2831-0b09-879e-68bb618d6d6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.778555] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Releasing lock "refresh_cache-a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.906330] env[62552]: DEBUG nova.scheduler.client.report [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 992.917506] env[62552]: DEBUG nova.network.neutron [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Successfully created port: 9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 993.034051] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 993.034541] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86956585-3317-4a43-a431-4a1b8a3d1439 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.052024] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 993.052024] env[62552]: value = "task-1239890" [ 993.052024] env[62552]: _type = "Task" [ 993.052024] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.063340] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239890, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.083460] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239886, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.131651] env[62552]: DEBUG oslo_concurrency.lockutils [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] Releasing lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.132081] env[62552]: DEBUG nova.compute.manager [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received event network-changed-34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 993.132413] env[62552]: DEBUG nova.compute.manager [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing instance network info cache due to event network-changed-34124ec0-5d30-4c38-8095-b11bda9f58fa. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 993.132652] env[62552]: DEBUG oslo_concurrency.lockutils [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] Acquiring lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.132886] env[62552]: DEBUG oslo_concurrency.lockutils [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] Acquired lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.133136] env[62552]: DEBUG nova.network.neutron [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing network info cache for port 34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.227728] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239889, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.243045] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528ac194-2831-0b09-879e-68bb618d6d6d, 'name': SearchDatastore_Task, 'duration_secs': 0.010242} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.243045] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.243045] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 505d7f8e-c27f-487e-98d8-c840a526d40e/505d7f8e-c27f-487e-98d8-c840a526d40e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 993.243523] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6dc9b44b-667d-4f63-ac95-38a0282c1b7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.253217] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 993.253217] env[62552]: value = "task-1239891" [ 993.253217] env[62552]: _type = "Task" [ 993.253217] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.264671] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.339979] env[62552]: DEBUG nova.compute.manager [req-4ce69692-6947-4b13-8e23-b427ff7045fe req-9c5459c8-2723-4e6b-8a74-5c5a8d9e84ef service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Received event network-vif-deleted-04e39cd5-39ae-4671-b715-e300dec6021d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 993.340218] env[62552]: INFO nova.compute.manager [req-4ce69692-6947-4b13-8e23-b427ff7045fe req-9c5459c8-2723-4e6b-8a74-5c5a8d9e84ef service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Neutron deleted interface 04e39cd5-39ae-4671-b715-e300dec6021d; detaching it from the instance and deleting it from the info cache [ 993.340451] env[62552]: DEBUG nova.network.neutron [req-4ce69692-6947-4b13-8e23-b427ff7045fe req-9c5459c8-2723-4e6b-8a74-5c5a8d9e84ef service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.423924] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.896s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.424886] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.877s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.427049] env[62552]: INFO nova.compute.claims [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.465086] env[62552]: INFO nova.scheduler.client.report [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted allocations for instance 304ad9f2-1965-4a70-aec9-0d2c931b8d95 [ 993.554728] env[62552]: DEBUG nova.compute.manager [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 993.574326] env[62552]: DEBUG oslo_vmware.api [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239890, 'name': PowerOffVM_Task, 'duration_secs': 0.231297} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.578523] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.578785] env[62552]: DEBUG nova.compute.manager [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 993.581830] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1d3740-4c33-42b5-98ba-ce0c1ebc0ce5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.597265] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.597406] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.597614] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.597840] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.598603] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.598603] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.598603] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.598603] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.598794] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.598966] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.599202] env[62552]: DEBUG nova.virt.hardware [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.603643] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b81a7bf-8c05-4fb2-8147-bb6dd1fbeeb9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.607051] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239886, 'name': Rename_Task, 'duration_secs': 1.127219} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.607974] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.608430] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2871002b-a708-4523-9025-9436c72ca273 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.619288] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edfd7aab-36d5-4cae-91a4-6e47cda5f35c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.624154] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 993.624154] env[62552]: value = "task-1239892" [ 993.624154] env[62552]: _type = "Task" [ 993.624154] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.649092] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239892, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.727745] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239889, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539409} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.732082] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9/0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 993.732444] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.732741] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9cea7480-e701-468d-a680-f2be655c972d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.739910] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 993.739910] env[62552]: value = "task-1239893" [ 993.739910] env[62552]: _type = "Task" [ 993.739910] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.750230] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239893, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.765110] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239891, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465639} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.765511] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 505d7f8e-c27f-487e-98d8-c840a526d40e/505d7f8e-c27f-487e-98d8-c840a526d40e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 993.765746] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.766013] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69644f09-b8c6-43bf-bdf7-d586be045193 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.768335] env[62552]: DEBUG nova.network.neutron [-] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.775586] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 993.775586] env[62552]: value = "task-1239894" [ 993.775586] env[62552]: _type = "Task" [ 993.775586] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.785390] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.785863] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239894, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.786140] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6e83f68-10ca-4632-a66f-a1faafdec75b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.794646] env[62552]: DEBUG oslo_vmware.api [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 993.794646] env[62552]: value = "task-1239895" [ 993.794646] env[62552]: _type = "Task" [ 993.794646] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.804100] env[62552]: DEBUG oslo_vmware.api [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239895, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.844899] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e533233-89fc-4573-b9e4-b96d094f9ec2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.879230] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defb9a75-79e2-43c5-a36e-bb0f1af1230b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.893884] env[62552]: DEBUG nova.compute.manager [req-4ce69692-6947-4b13-8e23-b427ff7045fe req-9c5459c8-2723-4e6b-8a74-5c5a8d9e84ef service nova] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Detach interface failed, port_id=04e39cd5-39ae-4671-b715-e300dec6021d, reason: Instance 04186fde-a9ed-415a-bd40-312f0347fcc7 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 993.974349] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b5d26129-9c3c-4d21-9ec6-0254b72499d1 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "304ad9f2-1965-4a70-aec9-0d2c931b8d95" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.831s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.112238] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.596s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.142924] env[62552]: DEBUG oslo_vmware.api [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239892, 'name': PowerOnVM_Task, 'duration_secs': 0.517334} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.143367] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.143607] env[62552]: INFO nova.compute.manager [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Took 9.66 seconds to spawn the instance on the hypervisor. [ 994.143796] env[62552]: DEBUG nova.compute.manager [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 994.144603] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f4152e-5f30-40d0-9e8b-e5048f6eef45 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.251497] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239893, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072267} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.251813] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.252626] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78663340-f708-4328-bfb5-3a00a96b71af {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.277127] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9/0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.277574] env[62552]: INFO nova.compute.manager [-] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Took 1.57 seconds to deallocate network for instance. [ 994.277857] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3dc3a0cb-755f-4ace-9376-ff911949a9f2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.305545] env[62552]: DEBUG nova.network.neutron [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updated VIF entry in instance network info cache for port 34124ec0-5d30-4c38-8095-b11bda9f58fa. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 994.306174] env[62552]: DEBUG nova.network.neutron [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [{"id": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "address": "fa:16:3e:c9:01:94", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34124ec0-5d", "ovs_interfaceid": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.314091] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 994.314091] env[62552]: value = "task-1239896" [ 994.314091] env[62552]: _type = "Task" [ 994.314091] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.314376] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239894, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07065} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.317104] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.319213] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c69f2c5-9d6b-4c2d-bc6d-2a7a4a58654c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.325663] env[62552]: DEBUG oslo_vmware.api [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239895, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.331963] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.350956] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 505d7f8e-c27f-487e-98d8-c840a526d40e/505d7f8e-c27f-487e-98d8-c840a526d40e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.351560] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61178e55-a9fc-46e5-bceb-40bb516be272 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.372376] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 994.372376] env[62552]: value = "task-1239897" [ 994.372376] env[62552]: _type = "Task" [ 994.372376] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.381226] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239897, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.627993] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.672318] env[62552]: INFO nova.compute.manager [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Took 20.16 seconds to build instance. [ 994.736395] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ee487e-3aaf-44b6-9f88-ff9196dc23be {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.746514] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4103c681-8bd3-4947-9b5a-793462d57f1c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.778357] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4a655e-acbb-4381-8642-aa55650282a9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.788331] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ac4aee-86c8-4bcd-b4e1-e2f318ea743c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.803845] env[62552]: DEBUG oslo_concurrency.lockutils [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.804441] env[62552]: DEBUG nova.compute.provider_tree [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.815786] env[62552]: DEBUG oslo_concurrency.lockutils [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] Releasing lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.816141] env[62552]: DEBUG nova.compute.manager [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 994.816318] env[62552]: DEBUG nova.compute.manager [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing instance network info cache due to event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 994.816534] env[62552]: DEBUG oslo_concurrency.lockutils [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] Acquiring lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.816814] env[62552]: DEBUG oslo_concurrency.lockutils [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] Acquired lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.816990] env[62552]: DEBUG nova.network.neutron [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.818709] env[62552]: DEBUG oslo_vmware.api [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239895, 'name': PowerOnVM_Task, 'duration_secs': 0.859359} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.822790] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.823040] env[62552]: DEBUG nova.compute.manager [None req-7f645b7d-1f80-4df7-b9fa-32d6257358a7 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 994.824637] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1cab1fe-5196-420a-9785-119d9ac529c6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.840845] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.884216] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239897, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.948945] env[62552]: DEBUG nova.compute.manager [req-88661fdd-b8ef-43b3-b559-cd64d4996c4e req-d678fc3e-430b-491d-a6dd-04ad7eee241c service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Received event network-vif-plugged-9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 994.949292] env[62552]: DEBUG oslo_concurrency.lockutils [req-88661fdd-b8ef-43b3-b559-cd64d4996c4e req-d678fc3e-430b-491d-a6dd-04ad7eee241c service nova] Acquiring lock "38bf968c-e757-4b46-b069-f625f5041c84-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.949571] env[62552]: DEBUG oslo_concurrency.lockutils [req-88661fdd-b8ef-43b3-b559-cd64d4996c4e req-d678fc3e-430b-491d-a6dd-04ad7eee241c service nova] Lock "38bf968c-e757-4b46-b069-f625f5041c84-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.949684] env[62552]: DEBUG oslo_concurrency.lockutils [req-88661fdd-b8ef-43b3-b559-cd64d4996c4e req-d678fc3e-430b-491d-a6dd-04ad7eee241c service nova] Lock "38bf968c-e757-4b46-b069-f625f5041c84-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.950074] env[62552]: DEBUG nova.compute.manager [req-88661fdd-b8ef-43b3-b559-cd64d4996c4e req-d678fc3e-430b-491d-a6dd-04ad7eee241c service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] No waiting events found dispatching network-vif-plugged-9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 994.950355] env[62552]: WARNING nova.compute.manager [req-88661fdd-b8ef-43b3-b559-cd64d4996c4e req-d678fc3e-430b-491d-a6dd-04ad7eee241c service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Received unexpected event network-vif-plugged-9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 for instance with vm_state building and task_state spawning. [ 995.157468] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.157718] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.175142] env[62552]: DEBUG oslo_concurrency.lockutils [None req-855597c7-ccc2-4abc-8491-efe364ed2b01 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.675s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.310523] env[62552]: DEBUG nova.scheduler.client.report [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 995.331644] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239896, 'name': ReconfigVM_Task, 'duration_secs': 0.800932} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.332044] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9/0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.332698] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7cac4fdc-2bf7-455c-957f-50ef95bbb73d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.339343] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 995.339343] env[62552]: value = "task-1239898" [ 995.339343] env[62552]: _type = "Task" [ 995.339343] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.348770] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239898, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.389579] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239897, 'name': ReconfigVM_Task, 'duration_secs': 0.579292} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.389958] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 505d7f8e-c27f-487e-98d8-c840a526d40e/505d7f8e-c27f-487e-98d8-c840a526d40e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.390737] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a84396fa-144c-4bd3-8359-6a5660ba7b81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.400253] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 995.400253] env[62552]: value = "task-1239899" [ 995.400253] env[62552]: _type = "Task" [ 995.400253] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.413864] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239899, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.486273] env[62552]: DEBUG nova.network.neutron [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Successfully updated port: 9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 995.535389] env[62552]: DEBUG nova.compute.manager [req-6f5acf19-1185-406d-9225-73e9a19c4b9f req-dbbf4807-067e-4dc1-b648-d009ddc142ef service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Received event network-changed-9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 995.535597] env[62552]: DEBUG nova.compute.manager [req-6f5acf19-1185-406d-9225-73e9a19c4b9f req-dbbf4807-067e-4dc1-b648-d009ddc142ef service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Refreshing instance network info cache due to event network-changed-9cf23ecf-7022-40a4-9182-2ac0d24f6fe4. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 995.535835] env[62552]: DEBUG oslo_concurrency.lockutils [req-6f5acf19-1185-406d-9225-73e9a19c4b9f req-dbbf4807-067e-4dc1-b648-d009ddc142ef service nova] Acquiring lock "refresh_cache-38bf968c-e757-4b46-b069-f625f5041c84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.536019] env[62552]: DEBUG oslo_concurrency.lockutils [req-6f5acf19-1185-406d-9225-73e9a19c4b9f req-dbbf4807-067e-4dc1-b648-d009ddc142ef service nova] Acquired lock "refresh_cache-38bf968c-e757-4b46-b069-f625f5041c84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.536163] env[62552]: DEBUG nova.network.neutron [req-6f5acf19-1185-406d-9225-73e9a19c4b9f req-dbbf4807-067e-4dc1-b648-d009ddc142ef service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Refreshing network info cache for port 9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.624303] env[62552]: DEBUG oslo_concurrency.lockutils [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.624566] env[62552]: DEBUG oslo_concurrency.lockutils [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.624775] env[62552]: DEBUG oslo_concurrency.lockutils [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.624960] env[62552]: DEBUG oslo_concurrency.lockutils [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.625150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.627298] env[62552]: INFO nova.compute.manager [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Terminating instance [ 995.660415] env[62552]: DEBUG nova.compute.manager [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 995.741590] env[62552]: DEBUG nova.network.neutron [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updated VIF entry in instance network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.742313] env[62552]: DEBUG nova.network.neutron [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.815323] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.815857] env[62552]: DEBUG nova.compute.manager [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 995.818915] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.757s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.818915] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.818915] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 995.819169] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.266s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.820591] env[62552]: INFO nova.compute.claims [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.823610] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced9d88d-8e40-45f5-bac5-90275e9e8ca9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.832985] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b55e032-dcbe-43d6-8abc-45cc24b695de {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.850593] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239898, 'name': Rename_Task, 'duration_secs': 0.216953} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.859359] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 995.860487] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d780ef0d-6434-4ca5-8f87-e083c3162c90 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.863897] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0753540a-f4e3-46de-8343-0ad5bb7f76b8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.870898] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 995.870898] env[62552]: value = "task-1239900" [ 995.870898] env[62552]: _type = "Task" [ 995.870898] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.876337] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26c1760-0759-4f37-a19f-9cc528b38a16 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.884961] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239900, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.910255] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179331MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 995.910337] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.920160] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239899, 'name': Rename_Task, 'duration_secs': 0.207489} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.920450] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 995.920713] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbebfe47-7e13-46ef-bcdb-b9778e5078f3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.929841] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 995.929841] env[62552]: value = "task-1239901" [ 995.929841] env[62552]: _type = "Task" [ 995.929841] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.939406] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.989670] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "refresh_cache-38bf968c-e757-4b46-b069-f625f5041c84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.085977] env[62552]: DEBUG nova.network.neutron [req-6f5acf19-1185-406d-9225-73e9a19c4b9f req-dbbf4807-067e-4dc1-b648-d009ddc142ef service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 996.131596] env[62552]: DEBUG nova.compute.manager [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 996.131838] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.132740] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34461cd-b0b9-4eaf-975c-fc9036076ad7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.140265] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 996.140505] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f12ab73-4ba6-405e-b617-82e53322c816 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.182799] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.229311] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 996.229550] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 996.229750] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleting the datastore file [datastore2] ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.230013] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b134b879-3bc0-49ad-baad-dc84531e3afc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.236062] env[62552]: DEBUG oslo_vmware.api [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 996.236062] env[62552]: value = "task-1239903" [ 996.236062] env[62552]: _type = "Task" [ 996.236062] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.244757] env[62552]: DEBUG oslo_vmware.api [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239903, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.247431] env[62552]: DEBUG oslo_concurrency.lockutils [req-617adbbc-227c-49bb-9584-572891e1a21f req-85495775-3133-4ea9-a40a-16e72410cbea service nova] Releasing lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.248925] env[62552]: DEBUG nova.network.neutron [req-6f5acf19-1185-406d-9225-73e9a19c4b9f req-dbbf4807-067e-4dc1-b648-d009ddc142ef service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.324816] env[62552]: DEBUG nova.compute.utils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.326620] env[62552]: DEBUG nova.compute.manager [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 996.326811] env[62552]: DEBUG nova.network.neutron [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 996.387242] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239900, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.389164] env[62552]: DEBUG nova.policy [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f87fac620384ff28a187546d6d29e2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25030025e90243b290a7d90efa26ec79', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.439135] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239901, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.746564] env[62552]: DEBUG oslo_vmware.api [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239903, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124407} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.746978] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.747302] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 996.747969] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 996.747969] env[62552]: INFO nova.compute.manager [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Took 0.62 seconds to destroy the instance on the hypervisor. [ 996.748112] env[62552]: DEBUG oslo.service.loopingcall [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.748367] env[62552]: DEBUG nova.compute.manager [-] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 996.748522] env[62552]: DEBUG nova.network.neutron [-] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.752092] env[62552]: DEBUG oslo_concurrency.lockutils [req-6f5acf19-1185-406d-9225-73e9a19c4b9f req-dbbf4807-067e-4dc1-b648-d009ddc142ef service nova] Releasing lock "refresh_cache-38bf968c-e757-4b46-b069-f625f5041c84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.752691] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "refresh_cache-38bf968c-e757-4b46-b069-f625f5041c84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.752919] env[62552]: DEBUG nova.network.neutron [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 996.830520] env[62552]: DEBUG nova.compute.manager [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 996.883796] env[62552]: DEBUG oslo_vmware.api [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239900, 'name': PowerOnVM_Task, 'duration_secs': 0.883389} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.884257] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 996.884583] env[62552]: INFO nova.compute.manager [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Took 9.94 seconds to spawn the instance on the hypervisor. [ 996.884887] env[62552]: DEBUG nova.compute.manager [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 996.889068] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ad570f-21bb-4a0c-84df-cac4385008a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.939938] env[62552]: DEBUG oslo_vmware.api [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239901, 'name': PowerOnVM_Task, 'duration_secs': 0.831847} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.940477] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 996.940836] env[62552]: DEBUG nova.compute.manager [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 996.942482] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565c5154-f76f-450c-ae47-ba729e925a03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.085707] env[62552]: DEBUG nova.network.neutron [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Successfully created port: dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.141126] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c694958-91c0-4d6b-91db-edf21716467f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.149255] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51034742-8311-469c-84db-25fef2cb57da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.182149] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e25cc6-5e1e-43b8-9e99-f13b6fbd19a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.190610] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749f1d96-59c6-4a51-8427-f0b2f04a066c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.206182] env[62552]: DEBUG nova.compute.provider_tree [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.285489] env[62552]: DEBUG nova.network.neutron [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 997.418642] env[62552]: INFO nova.compute.manager [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Took 22.84 seconds to build instance. [ 997.433935] env[62552]: DEBUG nova.network.neutron [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Updating instance_info_cache with network_info: [{"id": "9cf23ecf-7022-40a4-9182-2ac0d24f6fe4", "address": "fa:16:3e:3a:8b:84", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cf23ecf-70", "ovs_interfaceid": "9cf23ecf-7022-40a4-9182-2ac0d24f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.467232] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.485807] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "interface-6a4573f0-7152-4462-bb72-58a45b2cbd97-3688687a-8e5c-4347-8972-724ebb0df317" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.486240] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-6a4573f0-7152-4462-bb72-58a45b2cbd97-3688687a-8e5c-4347-8972-724ebb0df317" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.486797] env[62552]: DEBUG nova.objects.instance [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'flavor' on Instance uuid 6a4573f0-7152-4462-bb72-58a45b2cbd97 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.639820] env[62552]: DEBUG nova.compute.manager [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 997.641795] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8efe1a-5842-4b32-850a-8200662349db {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.709041] env[62552]: DEBUG nova.scheduler.client.report [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 997.781538] env[62552]: DEBUG nova.network.neutron [-] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.802342] env[62552]: DEBUG nova.compute.manager [req-fdeb2aa0-a858-414a-ba31-b60eebd44300 req-75d53735-017f-4273-a9ff-d64c69cfde1e service nova] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Received event network-vif-deleted-a1557e27-81a5-4443-a4f2-15be22c30b3c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 997.843263] env[62552]: DEBUG nova.compute.manager [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 997.869038] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 997.869282] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 997.869450] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 997.869691] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 997.869884] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 997.870071] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 997.870295] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 997.870465] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 997.870659] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 997.870830] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 997.871012] env[62552]: DEBUG nova.virt.hardware [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.873033] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f6890c-d97f-4a49-a122-a9869eb48038 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.880305] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb707a08-6365-4e7b-9290-551b8149a02f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.922050] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb447075-3852-49d7-8890-2dc87fdb8d67 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.348s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.938079] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "refresh_cache-38bf968c-e757-4b46-b069-f625f5041c84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.938533] env[62552]: DEBUG nova.compute.manager [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Instance network_info: |[{"id": "9cf23ecf-7022-40a4-9182-2ac0d24f6fe4", "address": "fa:16:3e:3a:8b:84", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cf23ecf-70", "ovs_interfaceid": "9cf23ecf-7022-40a4-9182-2ac0d24f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 997.939105] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:8b:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9cf23ecf-7022-40a4-9182-2ac0d24f6fe4', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.947435] env[62552]: DEBUG oslo.service.loopingcall [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.947701] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 997.947953] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7046b257-ea76-4b6f-a2d5-89f996b31735 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.967637] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.967637] env[62552]: value = "task-1239904" [ 997.967637] env[62552]: _type = "Task" [ 997.967637] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.975408] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239904, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.097487] env[62552]: DEBUG nova.objects.instance [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'pci_requests' on Instance uuid 6a4573f0-7152-4462-bb72-58a45b2cbd97 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.156620] env[62552]: INFO nova.compute.manager [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] instance snapshotting [ 998.160166] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a69272-81fb-472e-96ed-9e6b2ba59e40 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.181220] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f6c47e-2b66-4909-8ecb-c98b19bfe3f4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.214503] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.215029] env[62552]: DEBUG nova.compute.manager [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 998.217644] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.590s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.217847] env[62552]: DEBUG nova.objects.instance [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62552) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 998.284226] env[62552]: INFO nova.compute.manager [-] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Took 1.54 seconds to deallocate network for instance. [ 998.479108] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239904, 'name': CreateVM_Task} progress is 25%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.600028] env[62552]: DEBUG nova.objects.base [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Object Instance<6a4573f0-7152-4462-bb72-58a45b2cbd97> lazy-loaded attributes: flavor,pci_requests {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 998.600370] env[62552]: DEBUG nova.network.neutron [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 998.671980] env[62552]: DEBUG nova.policy [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6515846d33e44bca942767f9e1d650b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2f57f72de0c4adc9754458f4e7040b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.692378] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 998.694473] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b3b1dd4c-ea3c-48c4-b868-48f8171bf73a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.704614] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 998.704614] env[62552]: value = "task-1239905" [ 998.704614] env[62552]: _type = "Task" [ 998.704614] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.713859] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239905, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.722611] env[62552]: DEBUG nova.compute.utils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 998.728041] env[62552]: DEBUG nova.compute.manager [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 998.728041] env[62552]: DEBUG nova.network.neutron [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 998.791597] env[62552]: DEBUG oslo_concurrency.lockutils [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.855694] env[62552]: DEBUG nova.policy [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c732a4a15f9b4b16b241c09f4ba373e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3daa5253a0cc4538903b7cabc2f98fa2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.979546] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239904, 'name': CreateVM_Task, 'duration_secs': 0.78455} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.979755] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 998.980490] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.980701] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.981100] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 998.981415] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-760e4e1d-bf08-4cdc-94ac-777c5bb97cce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.986958] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 998.986958] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523ff5e1-f3f0-f524-7a09-59b77f6f7855" [ 998.986958] env[62552]: _type = "Task" [ 998.986958] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.997228] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523ff5e1-f3f0-f524-7a09-59b77f6f7855, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.048653] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "79166f2c-f864-4d8a-b1dc-e176710400cb" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.048653] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.201865] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "505d7f8e-c27f-487e-98d8-c840a526d40e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.202626] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "505d7f8e-c27f-487e-98d8-c840a526d40e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.202928] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "505d7f8e-c27f-487e-98d8-c840a526d40e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.203516] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "505d7f8e-c27f-487e-98d8-c840a526d40e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.203710] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "505d7f8e-c27f-487e-98d8-c840a526d40e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.209691] env[62552]: INFO nova.compute.manager [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Terminating instance [ 999.216245] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239905, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.230142] env[62552]: DEBUG nova.compute.manager [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 999.237025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b3bdcc1c-d327-4c08-99de-dacb1a56ccf0 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.237025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.432s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.237025] env[62552]: DEBUG oslo_concurrency.lockutils [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.238113] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.328s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.260750] env[62552]: INFO nova.scheduler.client.report [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted allocations for instance 04186fde-a9ed-415a-bd40-312f0347fcc7 [ 999.420562] env[62552]: DEBUG nova.network.neutron [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Successfully created port: d05003ff-4c67-4508-a70a-debf5ecec19e {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 999.498342] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523ff5e1-f3f0-f524-7a09-59b77f6f7855, 'name': SearchDatastore_Task, 'duration_secs': 0.012069} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.498677] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.498929] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 999.499182] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.499336] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.499529] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 999.499850] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1628d88-a230-4026-96cc-8ebb8615f0c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.508510] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 999.508753] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 999.509622] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0bfc74c-1c00-4eb4-aeef-b0b2413c846d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.515150] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 999.515150] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cc7be9-f758-0028-217f-a8377660db48" [ 999.515150] env[62552]: _type = "Task" [ 999.515150] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.523339] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cc7be9-f758-0028-217f-a8377660db48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.550672] env[62552]: DEBUG nova.compute.utils [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 999.580079] env[62552]: DEBUG nova.network.neutron [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Successfully updated port: dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.718853] env[62552]: DEBUG nova.compute.manager [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 999.719225] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.719562] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239905, 'name': CreateSnapshot_Task, 'duration_secs': 0.557051} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.720401] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ca6735-35fa-4376-86c1-e8888e9095cd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.723646] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 999.724746] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c9ae06-d793-400e-bfb3-d0c60908f1ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.745429] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.761921] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dec6eb42-426e-4ff9-a69e-cb3215c5463f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.770021] env[62552]: DEBUG oslo_vmware.api [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 999.770021] env[62552]: value = "task-1239906" [ 999.770021] env[62552]: _type = "Task" [ 999.770021] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.772425] env[62552]: DEBUG oslo_concurrency.lockutils [None req-511e6032-6d29-4076-83e0-6c8376f56ac0 tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "04186fde-a9ed-415a-bd40-312f0347fcc7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.064s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.787361] env[62552]: DEBUG oslo_vmware.api [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239906, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.027013] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cc7be9-f758-0028-217f-a8377660db48, 'name': SearchDatastore_Task, 'duration_secs': 0.00874} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.027864] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d681bd15-63b2-4488-b3e1-289f9a638f51 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.033532] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1000.033532] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5239e04c-0c1c-89c0-abd1-7c52d366c68b" [ 1000.033532] env[62552]: _type = "Task" [ 1000.033532] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.043292] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5239e04c-0c1c-89c0-abd1-7c52d366c68b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.054031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.082856] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.083045] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.083219] env[62552]: DEBUG nova.network.neutron [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.186945] env[62552]: DEBUG nova.network.neutron [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Successfully updated port: 3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1000.253054] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1000.254016] env[62552]: DEBUG nova.compute.manager [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1000.256277] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a40cdc8c-1345-4f18-9444-e8c2c9c57c6c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.267534] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1000.267534] env[62552]: value = "task-1239907" [ 1000.267534] env[62552]: _type = "Task" [ 1000.267534] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.278987] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239907, 'name': CloneVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.282460] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.282693] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.282889] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.283101] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.283262] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.283415] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.283622] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.283785] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.283954] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.284134] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.284311] env[62552]: DEBUG nova.virt.hardware [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.285074] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69d5ad8-2989-4f46-9f89-f52c55b8ffb1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.290397] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 88d0f502-fc3f-429e-bdf6-a1ebacec2117 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.290535] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.290656] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 11831580-1b58-476a-91ce-a4e55947fd91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.290774] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6c8b5367-4fa0-479b-9382-ff261201e3ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.290889] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6a5b5f4e-0c84-447d-a3da-8258512abb7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.290999] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 79166f2c-f864-4d8a-b1dc-e176710400cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.291128] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6a4573f0-7152-4462-bb72-58a45b2cbd97 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.291256] env[62552]: WARNING nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1000.291367] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.291474] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 505d7f8e-c27f-487e-98d8-c840a526d40e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.291580] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance d0aa144e-9bcf-4faa-8d4c-7743118fbe69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.291685] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.291824] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.291941] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 38bf968c-e757-4b46-b069-f625f5041c84 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.292065] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 158d4227-90ca-41ae-821b-efd353928cb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.292248] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 16785811-1927-408b-9ead-5848a5ffdf24 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.293642] env[62552]: DEBUG oslo_vmware.api [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239906, 'name': PowerOffVM_Task, 'duration_secs': 0.353277} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.294422] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.294603] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.294864] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-01e5d4e6-4b8f-4600-9abf-d23a9ee35ae7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.299622] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d45b11-0439-4113-8880-5170c33382e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.378840] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.379073] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.379225] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleting the datastore file [datastore1] 505d7f8e-c27f-487e-98d8-c840a526d40e {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.379492] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3637a94-9b2e-44f0-ac1c-4cac5d07f95f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.385816] env[62552]: DEBUG oslo_vmware.api [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1000.385816] env[62552]: value = "task-1239909" [ 1000.385816] env[62552]: _type = "Task" [ 1000.385816] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.394163] env[62552]: DEBUG oslo_vmware.api [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239909, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.544924] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5239e04c-0c1c-89c0-abd1-7c52d366c68b, 'name': SearchDatastore_Task, 'duration_secs': 0.011792} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.545225] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.545487] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 38bf968c-e757-4b46-b069-f625f5041c84/38bf968c-e757-4b46-b069-f625f5041c84.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1000.545753] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1bfd1521-3ef7-413e-90b8-2c9ccf6df90b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.552569] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1000.552569] env[62552]: value = "task-1239910" [ 1000.552569] env[62552]: _type = "Task" [ 1000.552569] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.562028] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239910, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.616776] env[62552]: DEBUG nova.network.neutron [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1000.689305] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.689507] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.689716] env[62552]: DEBUG nova.network.neutron [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.768064] env[62552]: DEBUG nova.network.neutron [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updating instance_info_cache with network_info: [{"id": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "address": "fa:16:3e:38:6f:9c", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1028f0-d5", "ovs_interfaceid": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.780824] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239907, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.795658] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1000.796068] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 15 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1000.796242] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3392MB phys_disk=200GB used_disk=15GB total_vcpus=48 used_vcpus=15 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1000.898290] env[62552]: DEBUG oslo_vmware.api [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239909, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13392} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.901571] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.901722] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.901971] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.902189] env[62552]: INFO nova.compute.manager [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1000.902459] env[62552]: DEBUG oslo.service.loopingcall [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.902927] env[62552]: DEBUG nova.compute.manager [-] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1000.903508] env[62552]: DEBUG nova.network.neutron [-] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1001.064869] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239910, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447134} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.066355] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 38bf968c-e757-4b46-b069-f625f5041c84/38bf968c-e757-4b46-b069-f625f5041c84.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1001.066584] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1001.067304] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c369231-190f-4509-a4b1-3044c0faf0ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.070252] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90dcd80d-f18c-4979-bff7-5c02de3f082d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.076553] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ad6b99-0850-4609-a5ee-33c18a6b3c55 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.080565] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1001.080565] env[62552]: value = "task-1239911" [ 1001.080565] env[62552]: _type = "Task" [ 1001.080565] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.110406] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb66922a-60f2-4b8a-bfb7-33a8056a15e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.116039] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239911, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.120988] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb72a9d-f07f-432f-bb92-a2026f891b8d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.125552] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "79166f2c-f864-4d8a-b1dc-e176710400cb" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.125800] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.126038] env[62552]: INFO nova.compute.manager [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Attaching volume 418c407e-7b74-49e2-8014-086b46b72f2a to /dev/sdb [ 1001.138162] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.176433] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4434958b-8965-4f39-96af-7c0d0c8988ad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.184614] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79334e85-ada3-41c1-9f90-c816ea0b3286 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.198396] env[62552]: DEBUG nova.virt.block_device [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating existing volume attachment record: 0a364fc1-bd89-4fe5-8a09-0b06b0c638a7 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1001.238488] env[62552]: WARNING nova.network.neutron [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] 06f1ea1a-5103-419f-a356-3b65b9a2685e already exists in list: networks containing: ['06f1ea1a-5103-419f-a356-3b65b9a2685e']. ignoring it [ 1001.275337] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.275337] env[62552]: DEBUG nova.compute.manager [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Instance network_info: |[{"id": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "address": "fa:16:3e:38:6f:9c", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1028f0-d5", "ovs_interfaceid": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1001.281159] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:6f:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd1028f0-d5a9-48be-84e3-9401cde0a1ff', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.287141] env[62552]: DEBUG oslo.service.loopingcall [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.288659] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239907, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.288901] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1001.289134] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a8d0762-a43b-4b84-ab40-f52c99cfe5ac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.309170] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.309170] env[62552]: value = "task-1239912" [ 1001.309170] env[62552]: _type = "Task" [ 1001.309170] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.320138] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239912, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.332609] env[62552]: DEBUG nova.network.neutron [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Successfully updated port: d05003ff-4c67-4508-a70a-debf5ecec19e {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1001.457551] env[62552]: DEBUG nova.compute.manager [req-703370c3-121a-44c0-9521-0d5c54855f00 req-ea7d7279-a2a2-4a9b-92eb-6ed18fd37f5e service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received event network-vif-plugged-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1001.457551] env[62552]: DEBUG oslo_concurrency.lockutils [req-703370c3-121a-44c0-9521-0d5c54855f00 req-ea7d7279-a2a2-4a9b-92eb-6ed18fd37f5e service nova] Acquiring lock "158d4227-90ca-41ae-821b-efd353928cb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.457551] env[62552]: DEBUG oslo_concurrency.lockutils [req-703370c3-121a-44c0-9521-0d5c54855f00 req-ea7d7279-a2a2-4a9b-92eb-6ed18fd37f5e service nova] Lock "158d4227-90ca-41ae-821b-efd353928cb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.457669] env[62552]: DEBUG oslo_concurrency.lockutils [req-703370c3-121a-44c0-9521-0d5c54855f00 req-ea7d7279-a2a2-4a9b-92eb-6ed18fd37f5e service nova] Lock "158d4227-90ca-41ae-821b-efd353928cb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.457788] env[62552]: DEBUG nova.compute.manager [req-703370c3-121a-44c0-9521-0d5c54855f00 req-ea7d7279-a2a2-4a9b-92eb-6ed18fd37f5e service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] No waiting events found dispatching network-vif-plugged-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1001.457982] env[62552]: WARNING nova.compute.manager [req-703370c3-121a-44c0-9521-0d5c54855f00 req-ea7d7279-a2a2-4a9b-92eb-6ed18fd37f5e service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received unexpected event network-vif-plugged-dd1028f0-d5a9-48be-84e3-9401cde0a1ff for instance with vm_state building and task_state spawning. [ 1001.539320] env[62552]: DEBUG nova.compute.manager [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Received event network-changed-13ae1de6-6606-415b-9afd-0bce65b48717 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1001.539517] env[62552]: DEBUG nova.compute.manager [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Refreshing instance network info cache due to event network-changed-13ae1de6-6606-415b-9afd-0bce65b48717. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1001.539729] env[62552]: DEBUG oslo_concurrency.lockutils [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] Acquiring lock "refresh_cache-0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.539883] env[62552]: DEBUG oslo_concurrency.lockutils [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] Acquired lock "refresh_cache-0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.540082] env[62552]: DEBUG nova.network.neutron [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Refreshing network info cache for port 13ae1de6-6606-415b-9afd-0bce65b48717 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1001.591615] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239911, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.13143} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.592256] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.593081] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc0cc3a-e659-4690-ba07-6027c817d271 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.638123] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 38bf968c-e757-4b46-b069-f625f5041c84/38bf968c-e757-4b46-b069-f625f5041c84.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.641441] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3da89b5d-c596-47ca-8971-ce05ce30bce9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.664505] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1001.676803] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1001.676803] env[62552]: value = "task-1239916" [ 1001.676803] env[62552]: _type = "Task" [ 1001.676803] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.687725] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239916, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.718973] env[62552]: DEBUG nova.network.neutron [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3688687a-8e5c-4347-8972-724ebb0df317", "address": "fa:16:3e:79:0b:d3", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3688687a-8e", "ovs_interfaceid": "3688687a-8e5c-4347-8972-724ebb0df317", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.731483] env[62552]: DEBUG nova.compute.manager [req-9df0e3e6-2b29-4c97-9dfb-49ee7e37944b req-7bdbab13-bc3b-4fa7-937e-bf6a94a589ee service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received event network-vif-plugged-3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1001.731725] env[62552]: DEBUG oslo_concurrency.lockutils [req-9df0e3e6-2b29-4c97-9dfb-49ee7e37944b req-7bdbab13-bc3b-4fa7-937e-bf6a94a589ee service nova] Acquiring lock "6a4573f0-7152-4462-bb72-58a45b2cbd97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.732080] env[62552]: DEBUG oslo_concurrency.lockutils [req-9df0e3e6-2b29-4c97-9dfb-49ee7e37944b req-7bdbab13-bc3b-4fa7-937e-bf6a94a589ee service nova] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.732208] env[62552]: DEBUG oslo_concurrency.lockutils [req-9df0e3e6-2b29-4c97-9dfb-49ee7e37944b req-7bdbab13-bc3b-4fa7-937e-bf6a94a589ee service nova] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.732383] env[62552]: DEBUG nova.compute.manager [req-9df0e3e6-2b29-4c97-9dfb-49ee7e37944b req-7bdbab13-bc3b-4fa7-937e-bf6a94a589ee service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] No waiting events found dispatching network-vif-plugged-3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1001.732555] env[62552]: WARNING nova.compute.manager [req-9df0e3e6-2b29-4c97-9dfb-49ee7e37944b req-7bdbab13-bc3b-4fa7-937e-bf6a94a589ee service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received unexpected event network-vif-plugged-3688687a-8e5c-4347-8972-724ebb0df317 for instance with vm_state active and task_state None. [ 1001.770577] env[62552]: DEBUG nova.network.neutron [-] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.781382] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239907, 'name': CloneVM_Task, 'duration_secs': 1.437506} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.781693] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Created linked-clone VM from snapshot [ 1001.782527] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e837812a-ce34-491b-bd59-1cff14d40d10 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.790532] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Uploading image b717a313-6a55-4657-b3ae-5bbf16e197c7 {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1001.803127] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1001.803493] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-74d872c9-9482-43ac-871d-a9241af812f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.811075] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1001.811075] env[62552]: value = "task-1239917" [ 1001.811075] env[62552]: _type = "Task" [ 1001.811075] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.822952] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239912, 'name': CreateVM_Task, 'duration_secs': 0.390325} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.825487] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1001.825800] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239917, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.826486] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.826672] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.827013] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.827292] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaa08cac-bed5-46b9-9a20-29045937030c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.832376] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1001.832376] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a93ace-f26e-f8e0-c519-011d98ecf9c4" [ 1001.832376] env[62552]: _type = "Task" [ 1001.832376] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.836117] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "refresh_cache-16785811-1927-408b-9ead-5848a5ffdf24" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.837998] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "refresh_cache-16785811-1927-408b-9ead-5848a5ffdf24" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.837998] env[62552]: DEBUG nova.network.neutron [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.843049] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a93ace-f26e-f8e0-c519-011d98ecf9c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.170831] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1002.171204] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.933s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.171465] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.989s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.173149] env[62552]: INFO nova.compute.claims [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.186511] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239916, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.222785] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.223318] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.223506] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.224356] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fefb076-6422-4ae9-a9cc-a62c3e87d059 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.243104] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1002.243413] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1002.243536] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.243719] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1002.243866] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.244054] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1002.244276] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1002.244435] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1002.244601] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1002.244765] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1002.244950] env[62552]: DEBUG nova.virt.hardware [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.251333] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Reconfiguring VM to attach interface {{(pid=62552) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1002.251958] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53cfd020-cd8a-4c54-aaf5-1717b1824a78 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.269638] env[62552]: DEBUG oslo_vmware.api [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 1002.269638] env[62552]: value = "task-1239918" [ 1002.269638] env[62552]: _type = "Task" [ 1002.269638] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.278192] env[62552]: INFO nova.compute.manager [-] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Took 1.38 seconds to deallocate network for instance. [ 1002.278482] env[62552]: DEBUG oslo_vmware.api [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239918, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.303844] env[62552]: DEBUG nova.network.neutron [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Updated VIF entry in instance network info cache for port 13ae1de6-6606-415b-9afd-0bce65b48717. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1002.304245] env[62552]: DEBUG nova.network.neutron [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Updating instance_info_cache with network_info: [{"id": "13ae1de6-6606-415b-9afd-0bce65b48717", "address": "fa:16:3e:b8:4a:1f", "network": {"id": "5b1bbdf9-0cfe-49da-bcf5-4f8b41744f19", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-106952979-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f44eb24ac74742bdbc1ce860fcfde584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13ae1de6-66", "ovs_interfaceid": "13ae1de6-6606-415b-9afd-0bce65b48717", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.327091] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239917, 'name': Destroy_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.344106] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a93ace-f26e-f8e0-c519-011d98ecf9c4, 'name': SearchDatastore_Task, 'duration_secs': 0.009749} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.344427] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.344669] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.344910] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.345074] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.345261] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.345880] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81563677-b16b-40ab-acfb-989392c72b2a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.354212] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.354413] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1002.355202] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31feeffa-3932-4624-8460-2f4c7b34e17d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.360475] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1002.360475] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526e3af0-e13f-5277-04f2-727960dace65" [ 1002.360475] env[62552]: _type = "Task" [ 1002.360475] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.368102] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526e3af0-e13f-5277-04f2-727960dace65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.372941] env[62552]: DEBUG nova.network.neutron [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1002.538962] env[62552]: DEBUG nova.network.neutron [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Updating instance_info_cache with network_info: [{"id": "d05003ff-4c67-4508-a70a-debf5ecec19e", "address": "fa:16:3e:79:18:5b", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd05003ff-4c", "ovs_interfaceid": "d05003ff-4c67-4508-a70a-debf5ecec19e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.689079] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239916, 'name': ReconfigVM_Task, 'duration_secs': 0.656238} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.689931] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 38bf968c-e757-4b46-b069-f625f5041c84/38bf968c-e757-4b46-b069-f625f5041c84.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.690557] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4dfabedc-1845-4a56-900e-5b22a1a7967b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.696944] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1002.696944] env[62552]: value = "task-1239919" [ 1002.696944] env[62552]: _type = "Task" [ 1002.696944] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.706296] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239919, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.781418] env[62552]: DEBUG oslo_vmware.api [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239918, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.784763] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.806614] env[62552]: DEBUG oslo_concurrency.lockutils [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] Releasing lock "refresh_cache-0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.806919] env[62552]: DEBUG nova.compute.manager [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received event network-changed-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1002.807181] env[62552]: DEBUG nova.compute.manager [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Refreshing instance network info cache due to event network-changed-dd1028f0-d5a9-48be-84e3-9401cde0a1ff. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1002.807402] env[62552]: DEBUG oslo_concurrency.lockutils [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] Acquiring lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.807550] env[62552]: DEBUG oslo_concurrency.lockutils [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] Acquired lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.807719] env[62552]: DEBUG nova.network.neutron [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Refreshing network info cache for port dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.826256] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239917, 'name': Destroy_Task, 'duration_secs': 0.690896} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.826525] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Destroyed the VM [ 1002.826766] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1002.827303] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3c8e4369-e4bc-4a41-b140-aa4cf35ef3c4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.833013] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1002.833013] env[62552]: value = "task-1239920" [ 1002.833013] env[62552]: _type = "Task" [ 1002.833013] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.840992] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239920, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.869967] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526e3af0-e13f-5277-04f2-727960dace65, 'name': SearchDatastore_Task, 'duration_secs': 0.011713} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.870849] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dbfdb4c-8bec-4291-bb34-edf2eb8461e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.876151] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1002.876151] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521f24da-292e-15ae-4628-20cffa76ed07" [ 1002.876151] env[62552]: _type = "Task" [ 1002.876151] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.886661] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521f24da-292e-15ae-4628-20cffa76ed07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.042274] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "refresh_cache-16785811-1927-408b-9ead-5848a5ffdf24" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.042637] env[62552]: DEBUG nova.compute.manager [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Instance network_info: |[{"id": "d05003ff-4c67-4508-a70a-debf5ecec19e", "address": "fa:16:3e:79:18:5b", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd05003ff-4c", "ovs_interfaceid": "d05003ff-4c67-4508-a70a-debf5ecec19e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1003.043215] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:18:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e41c97-4d75-4041-ae71-321e7e9d480b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd05003ff-4c67-4508-a70a-debf5ecec19e', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1003.052799] env[62552]: DEBUG oslo.service.loopingcall [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.053921] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1003.053921] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e3a909a-1aaf-4422-8ffb-cf0f11e2a6ec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.074764] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1003.074764] env[62552]: value = "task-1239921" [ 1003.074764] env[62552]: _type = "Task" [ 1003.074764] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.082619] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239921, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.210033] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239919, 'name': Rename_Task, 'duration_secs': 0.239752} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.210496] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1003.210784] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-426467eb-d17f-4707-a741-05f972edebf3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.217351] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1003.217351] env[62552]: value = "task-1239922" [ 1003.217351] env[62552]: _type = "Task" [ 1003.217351] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.228728] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239922, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.284396] env[62552]: DEBUG oslo_vmware.api [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239918, 'name': ReconfigVM_Task, 'duration_secs': 0.563073} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.284942] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.285187] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Reconfigured VM to attach interface {{(pid=62552) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1003.343323] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239920, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.394777] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521f24da-292e-15ae-4628-20cffa76ed07, 'name': SearchDatastore_Task, 'duration_secs': 0.010367} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.398220] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.398706] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2/158d4227-90ca-41ae-821b-efd353928cb2.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1003.399487] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efc222a9-e91d-472e-ae6d-59ea8d1052d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.406158] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1003.406158] env[62552]: value = "task-1239923" [ 1003.406158] env[62552]: _type = "Task" [ 1003.406158] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.416258] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239923, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.434417] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01e5973-f85a-4286-9392-9ee8177ea3eb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.442474] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63e869d-f69c-4138-8898-6cdfb349d32f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.478343] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52d6d30-2475-4522-9fb6-da112faa0c60 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.486430] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec221b2-e4bf-4bbf-93e6-65df5685adbc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.501538] env[62552]: DEBUG nova.compute.provider_tree [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.557928] env[62552]: DEBUG nova.network.neutron [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updated VIF entry in instance network info cache for port dd1028f0-d5a9-48be-84e3-9401cde0a1ff. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1003.558445] env[62552]: DEBUG nova.network.neutron [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updating instance_info_cache with network_info: [{"id": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "address": "fa:16:3e:38:6f:9c", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1028f0-d5", "ovs_interfaceid": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.568966] env[62552]: DEBUG nova.compute.manager [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Received event network-vif-plugged-d05003ff-4c67-4508-a70a-debf5ecec19e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1003.569299] env[62552]: DEBUG oslo_concurrency.lockutils [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] Acquiring lock "16785811-1927-408b-9ead-5848a5ffdf24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.569541] env[62552]: DEBUG oslo_concurrency.lockutils [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] Lock "16785811-1927-408b-9ead-5848a5ffdf24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.569680] env[62552]: DEBUG oslo_concurrency.lockutils [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] Lock "16785811-1927-408b-9ead-5848a5ffdf24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.569911] env[62552]: DEBUG nova.compute.manager [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] No waiting events found dispatching network-vif-plugged-d05003ff-4c67-4508-a70a-debf5ecec19e {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1003.570163] env[62552]: WARNING nova.compute.manager [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Received unexpected event network-vif-plugged-d05003ff-4c67-4508-a70a-debf5ecec19e for instance with vm_state building and task_state spawning. [ 1003.570351] env[62552]: DEBUG nova.compute.manager [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Received event network-changed-d05003ff-4c67-4508-a70a-debf5ecec19e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1003.570528] env[62552]: DEBUG nova.compute.manager [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Refreshing instance network info cache due to event network-changed-d05003ff-4c67-4508-a70a-debf5ecec19e. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1003.570732] env[62552]: DEBUG oslo_concurrency.lockutils [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] Acquiring lock "refresh_cache-16785811-1927-408b-9ead-5848a5ffdf24" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.570980] env[62552]: DEBUG oslo_concurrency.lockutils [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] Acquired lock "refresh_cache-16785811-1927-408b-9ead-5848a5ffdf24" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.571092] env[62552]: DEBUG nova.network.neutron [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Refreshing network info cache for port d05003ff-4c67-4508-a70a-debf5ecec19e {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1003.584980] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239921, 'name': CreateVM_Task, 'duration_secs': 0.444175} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.585174] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1003.585875] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.586061] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.586412] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.586921] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1bc0edb-0a4d-4a86-abd4-59bb843b92e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.591600] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1003.591600] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52659eaf-7043-59b9-523f-463a612bfdba" [ 1003.591600] env[62552]: _type = "Task" [ 1003.591600] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.602816] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52659eaf-7043-59b9-523f-463a612bfdba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.728433] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239922, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.785500] env[62552]: DEBUG nova.compute.manager [req-ed4c287e-416a-47fa-a2b6-dc4b99ee417b req-4431086d-2d07-44f0-aee9-5e43a3481fa0 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received event network-changed-3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1003.786035] env[62552]: DEBUG nova.compute.manager [req-ed4c287e-416a-47fa-a2b6-dc4b99ee417b req-4431086d-2d07-44f0-aee9-5e43a3481fa0 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing instance network info cache due to event network-changed-3688687a-8e5c-4347-8972-724ebb0df317. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1003.786272] env[62552]: DEBUG oslo_concurrency.lockutils [req-ed4c287e-416a-47fa-a2b6-dc4b99ee417b req-4431086d-2d07-44f0-aee9-5e43a3481fa0 service nova] Acquiring lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.786521] env[62552]: DEBUG oslo_concurrency.lockutils [req-ed4c287e-416a-47fa-a2b6-dc4b99ee417b req-4431086d-2d07-44f0-aee9-5e43a3481fa0 service nova] Acquired lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.786787] env[62552]: DEBUG nova.network.neutron [req-ed4c287e-416a-47fa-a2b6-dc4b99ee417b req-4431086d-2d07-44f0-aee9-5e43a3481fa0 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing network info cache for port 3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1003.790055] env[62552]: DEBUG oslo_concurrency.lockutils [None req-95667d97-75cf-4926-afae-e5892208d312 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-6a4573f0-7152-4462-bb72-58a45b2cbd97-3688687a-8e5c-4347-8972-724ebb0df317" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.304s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.844353] env[62552]: DEBUG oslo_vmware.api [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239920, 'name': RemoveSnapshot_Task, 'duration_secs': 0.857862} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.844673] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1003.916646] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239923, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492511} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.916986] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2/158d4227-90ca-41ae-821b-efd353928cb2.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1003.917236] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.917490] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90170491-a481-469d-9f91-54221d3e0e00 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.925115] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1003.925115] env[62552]: value = "task-1239925" [ 1003.925115] env[62552]: _type = "Task" [ 1003.925115] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.932926] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239925, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.005186] env[62552]: DEBUG nova.scheduler.client.report [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1004.061621] env[62552]: DEBUG oslo_concurrency.lockutils [req-f20c4192-64b4-450b-af2a-e7e875c2b09d req-65272231-d1fe-48fa-814b-87434e9652cd service nova] Releasing lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.102253] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52659eaf-7043-59b9-523f-463a612bfdba, 'name': SearchDatastore_Task, 'duration_secs': 0.060344} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.102556] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.102810] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.103076] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.103234] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.103418] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.103670] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77d5077b-a5c0-4707-90d0-ae5ee5bc9a5f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.119549] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.119739] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1004.120496] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c318b0af-a1d4-4870-b91c-61fabf6ab3ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.125271] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1004.125271] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e202be-7b07-727c-ccfa-ad13b776e925" [ 1004.125271] env[62552]: _type = "Task" [ 1004.125271] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.132553] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e202be-7b07-727c-ccfa-ad13b776e925, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.227312] env[62552]: DEBUG oslo_vmware.api [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1239922, 'name': PowerOnVM_Task, 'duration_secs': 0.663877} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.227608] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1004.227784] env[62552]: INFO nova.compute.manager [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Took 10.67 seconds to spawn the instance on the hypervisor. [ 1004.227972] env[62552]: DEBUG nova.compute.manager [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1004.228743] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7473d4-0220-40d5-b35a-880ecca4e58a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.349762] env[62552]: WARNING nova.compute.manager [None req-124ec15d-bdd7-4df6-b2d6-dd8bd64eb9a4 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Image not found during snapshot: nova.exception.ImageNotFound: Image b717a313-6a55-4657-b3ae-5bbf16e197c7 could not be found. [ 1004.390884] env[62552]: DEBUG nova.network.neutron [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Updated VIF entry in instance network info cache for port d05003ff-4c67-4508-a70a-debf5ecec19e. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1004.391320] env[62552]: DEBUG nova.network.neutron [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Updating instance_info_cache with network_info: [{"id": "d05003ff-4c67-4508-a70a-debf5ecec19e", "address": "fa:16:3e:79:18:5b", "network": {"id": "7034b4cd-5699-4249-876c-a8792ccb4726", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-539405807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3daa5253a0cc4538903b7cabc2f98fa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd05003ff-4c", "ovs_interfaceid": "d05003ff-4c67-4508-a70a-debf5ecec19e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.435025] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239925, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069412} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.436094] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.436312] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc329294-d87e-4740-abae-795f64814572 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.457578] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2/158d4227-90ca-41ae-821b-efd353928cb2.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.459925] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d618c043-2e2e-4094-85f3-1383c632f958 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.479782] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1004.479782] env[62552]: value = "task-1239926" [ 1004.479782] env[62552]: _type = "Task" [ 1004.479782] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.489771] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239926, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.510349] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.510957] env[62552]: DEBUG nova.compute.manager [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1004.513878] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 7.047s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.514143] env[62552]: DEBUG nova.objects.instance [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62552) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1004.619012] env[62552]: DEBUG nova.network.neutron [req-ed4c287e-416a-47fa-a2b6-dc4b99ee417b req-4431086d-2d07-44f0-aee9-5e43a3481fa0 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updated VIF entry in instance network info cache for port 3688687a-8e5c-4347-8972-724ebb0df317. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1004.619481] env[62552]: DEBUG nova.network.neutron [req-ed4c287e-416a-47fa-a2b6-dc4b99ee417b req-4431086d-2d07-44f0-aee9-5e43a3481fa0 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3688687a-8e5c-4347-8972-724ebb0df317", "address": "fa:16:3e:79:0b:d3", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3688687a-8e", "ovs_interfaceid": "3688687a-8e5c-4347-8972-724ebb0df317", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.636313] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e202be-7b07-727c-ccfa-ad13b776e925, 'name': SearchDatastore_Task, 'duration_secs': 0.033988} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.637174] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aeb12a63-b955-402a-8943-eda1160444c3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.642639] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1004.642639] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dfe51c-dd81-2763-d09e-129aea2b4fd7" [ 1004.642639] env[62552]: _type = "Task" [ 1004.642639] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.653685] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dfe51c-dd81-2763-d09e-129aea2b4fd7, 'name': SearchDatastore_Task, 'duration_secs': 0.008355} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.653981] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.654297] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 16785811-1927-408b-9ead-5848a5ffdf24/16785811-1927-408b-9ead-5848a5ffdf24.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.654578] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9ccf721-1d3e-43db-ae84-5fa0f7acc759 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.661201] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1004.661201] env[62552]: value = "task-1239927" [ 1004.661201] env[62552]: _type = "Task" [ 1004.661201] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.669271] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239927, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.749397] env[62552]: INFO nova.compute.manager [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Took 21.20 seconds to build instance. [ 1004.894292] env[62552]: DEBUG oslo_concurrency.lockutils [req-43c678a8-0213-42f7-9088-fc408b5c52b7 req-85f4840e-ce7a-465d-a192-fd66e6abc01b service nova] Releasing lock "refresh_cache-16785811-1927-408b-9ead-5848a5ffdf24" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.991042] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239926, 'name': ReconfigVM_Task, 'duration_secs': 0.292965} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.991306] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2/158d4227-90ca-41ae-821b-efd353928cb2.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.992135] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-558fabc9-67d2-473e-b3ba-326e9b716971 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.000478] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1005.000478] env[62552]: value = "task-1239928" [ 1005.000478] env[62552]: _type = "Task" [ 1005.000478] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.009517] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239928, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.018200] env[62552]: DEBUG nova.compute.utils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.025029] env[62552]: DEBUG nova.compute.manager [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1005.025229] env[62552]: DEBUG nova.network.neutron [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1005.098446] env[62552]: DEBUG nova.policy [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7f8aa8200874dddb71d8b21bd12ca04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8b8d96b464a439e9c7ef6f3e419a9bc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1005.123069] env[62552]: DEBUG oslo_concurrency.lockutils [req-ed4c287e-416a-47fa-a2b6-dc4b99ee417b req-4431086d-2d07-44f0-aee9-5e43a3481fa0 service nova] Releasing lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.123069] env[62552]: DEBUG nova.compute.manager [req-ed4c287e-416a-47fa-a2b6-dc4b99ee417b req-4431086d-2d07-44f0-aee9-5e43a3481fa0 service nova] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Received event network-vif-deleted-7734d400-5df3-4f55-b6da-b52dfd148993 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1005.175023] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239927, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497245} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.175023] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 16785811-1927-408b-9ead-5848a5ffdf24/16785811-1927-408b-9ead-5848a5ffdf24.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.175023] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.177382] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b55d74ee-d95a-4138-a527-a55d447046d5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.184823] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1005.184823] env[62552]: value = "task-1239929" [ 1005.184823] env[62552]: _type = "Task" [ 1005.184823] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.196271] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239929, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.200480] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "interface-6a4573f0-7152-4462-bb72-58a45b2cbd97-3688687a-8e5c-4347-8972-724ebb0df317" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.200913] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-6a4573f0-7152-4462-bb72-58a45b2cbd97-3688687a-8e5c-4347-8972-724ebb0df317" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.251313] env[62552]: DEBUG oslo_concurrency.lockutils [None req-df5cff66-b941-4ea3-92a5-9407db7873fa tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.713s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.519019] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239928, 'name': Rename_Task, 'duration_secs': 0.188063} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.519019] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.519019] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ed6dad2-5db2-46f7-b5e0-1a7ffafccd37 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.525688] env[62552]: DEBUG nova.compute.manager [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1005.533416] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1005.533416] env[62552]: value = "task-1239930" [ 1005.533416] env[62552]: _type = "Task" [ 1005.533416] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.533416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-851b76a5-c051-42e9-8283-3655355d3ce2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.533416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.741s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.533416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.535129] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.750s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.535457] env[62552]: DEBUG nova.objects.instance [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'resources' on Instance uuid 505d7f8e-c27f-487e-98d8-c840a526d40e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.546974] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239930, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.568055] env[62552]: DEBUG nova.network.neutron [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Successfully created port: 85efda2a-fdf8-4118-861c-af85ee3f0d46 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1005.584543] env[62552]: INFO nova.scheduler.client.report [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted allocations for instance ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71 [ 1005.685030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.685030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.685030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.685030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.685030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.691956] env[62552]: INFO nova.compute.manager [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Terminating instance [ 1005.700020] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239929, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.187984} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.700480] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.701481] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4630071e-eee7-420f-a766-6ebb44f31bbc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.706521] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.706978] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.708270] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9974c94-696d-45ed-ad06-df4cbf9235e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.743024] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 16785811-1927-408b-9ead-5848a5ffdf24/16785811-1927-408b-9ead-5848a5ffdf24.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.758828] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-240261d6-27e9-48eb-8433-9af502729e32 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.779707] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1005.780008] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267504', 'volume_id': '418c407e-7b74-49e2-8014-086b46b72f2a', 'name': 'volume-418c407e-7b74-49e2-8014-086b46b72f2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '79166f2c-f864-4d8a-b1dc-e176710400cb', 'attached_at': '', 'detached_at': '', 'volume_id': '418c407e-7b74-49e2-8014-086b46b72f2a', 'serial': '418c407e-7b74-49e2-8014-086b46b72f2a'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1005.781242] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243bb8f8-a114-4108-bb6d-9a3edcada6f3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.784251] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b56fd0e-c602-4973-98a5-3ec31aaadd31 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.792234] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1005.792234] env[62552]: value = "task-1239931" [ 1005.792234] env[62552]: _type = "Task" [ 1005.792234] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.828633] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Reconfiguring VM to detach interface {{(pid=62552) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1005.829835] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-340d527a-a648-4311-8082-f065c2341c26 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.834866] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7bdc55c-315b-4c98-b549-1dd950533af6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.855039] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239931, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.878076] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] volume-418c407e-7b74-49e2-8014-086b46b72f2a/volume-418c407e-7b74-49e2-8014-086b46b72f2a.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.878552] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 1005.878552] env[62552]: value = "task-1239932" [ 1005.878552] env[62552]: _type = "Task" [ 1005.878552] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.878853] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca69d442-806e-4abd-8123-7be379c91ce7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.904986] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.906670] env[62552]: DEBUG oslo_vmware.api [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1005.906670] env[62552]: value = "task-1239933" [ 1005.906670] env[62552]: _type = "Task" [ 1005.906670] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.915167] env[62552]: DEBUG oslo_vmware.api [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239933, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.024465] env[62552]: DEBUG nova.compute.manager [req-0cb27016-1b07-4312-ac6c-7cff24f13558 req-fb665255-8743-44fe-9a43-303862bf370d service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Received event network-changed-9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1006.024766] env[62552]: DEBUG nova.compute.manager [req-0cb27016-1b07-4312-ac6c-7cff24f13558 req-fb665255-8743-44fe-9a43-303862bf370d service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Refreshing instance network info cache due to event network-changed-9cf23ecf-7022-40a4-9182-2ac0d24f6fe4. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1006.025146] env[62552]: DEBUG oslo_concurrency.lockutils [req-0cb27016-1b07-4312-ac6c-7cff24f13558 req-fb665255-8743-44fe-9a43-303862bf370d service nova] Acquiring lock "refresh_cache-38bf968c-e757-4b46-b069-f625f5041c84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.026017] env[62552]: DEBUG oslo_concurrency.lockutils [req-0cb27016-1b07-4312-ac6c-7cff24f13558 req-fb665255-8743-44fe-9a43-303862bf370d service nova] Acquired lock "refresh_cache-38bf968c-e757-4b46-b069-f625f5041c84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.026017] env[62552]: DEBUG nova.network.neutron [req-0cb27016-1b07-4312-ac6c-7cff24f13558 req-fb665255-8743-44fe-9a43-303862bf370d service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Refreshing network info cache for port 9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.055290] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239930, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.098033] env[62552]: DEBUG oslo_concurrency.lockutils [None req-006c0d70-2c45-4ce7-857c-9bf6eacf898a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.473s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.196520] env[62552]: DEBUG nova.compute.manager [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1006.196819] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1006.197869] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653edc6b-405a-440b-8ebb-ac0ec6e69eb3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.205738] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.206022] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8af2d101-4a1c-408f-84a1-cd6110d44682 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.211932] env[62552]: DEBUG oslo_vmware.api [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1006.211932] env[62552]: value = "task-1239934" [ 1006.211932] env[62552]: _type = "Task" [ 1006.211932] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.219032] env[62552]: DEBUG oslo_vmware.api [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239934, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.283091] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9b0a74-9f21-4ddc-b8d6-21ba22b1031f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.291445] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9599040c-bc44-4a89-b931-1f716b4b64e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.327240] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c9ef99-76e4-404e-b762-59f858c24a1c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.337599] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa99ac03-3073-420b-bdf0-21fe9478d961 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.341254] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239931, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.351382] env[62552]: DEBUG nova.compute.provider_tree [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.405259] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.414797] env[62552]: DEBUG oslo_vmware.api [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239933, 'name': ReconfigVM_Task, 'duration_secs': 0.4152} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.415069] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfigured VM instance instance-00000055 to attach disk [datastore2] volume-418c407e-7b74-49e2-8014-086b46b72f2a/volume-418c407e-7b74-49e2-8014-086b46b72f2a.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.419563] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9de9ff5-9a6f-4295-9b06-211ab4f18327 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.433809] env[62552]: DEBUG oslo_vmware.api [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1006.433809] env[62552]: value = "task-1239935" [ 1006.433809] env[62552]: _type = "Task" [ 1006.433809] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.442062] env[62552]: DEBUG oslo_vmware.api [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239935, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.542620] env[62552]: DEBUG oslo_vmware.api [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1239930, 'name': PowerOnVM_Task, 'duration_secs': 0.661549} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.542934] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.543169] env[62552]: INFO nova.compute.manager [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Took 8.70 seconds to spawn the instance on the hypervisor. [ 1006.543786] env[62552]: DEBUG nova.compute.manager [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1006.544512] env[62552]: DEBUG nova.compute.manager [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1006.546984] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f2cc75-748d-439f-9d86-2651b2a38f52 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.575945] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1006.576260] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1006.576423] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.576721] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1006.576997] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.577211] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1006.577440] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1006.577610] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1006.577785] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1006.577953] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1006.578155] env[62552]: DEBUG nova.virt.hardware [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.579085] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afbfafa-d124-4a54-a934-7d9403670042 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.592195] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e16eb1d-ea3b-4ce7-948b-800c4d6260cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.721781] env[62552]: DEBUG oslo_vmware.api [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239934, 'name': PowerOffVM_Task, 'duration_secs': 0.353014} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.724273] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1006.724472] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1006.724744] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9350b154-5e3f-445e-90e7-319c723e79c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.811145] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1006.811408] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1006.811578] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Deleting the datastore file [datastore2] c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1006.811923] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-311608e2-bcf0-4d60-a814-a945ad36441a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.820239] env[62552]: DEBUG oslo_vmware.api [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1006.820239] env[62552]: value = "task-1239937" [ 1006.820239] env[62552]: _type = "Task" [ 1006.820239] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.834308] env[62552]: DEBUG oslo_vmware.api [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239937, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.838372] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239931, 'name': ReconfigVM_Task, 'duration_secs': 0.748104} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.838372] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 16785811-1927-408b-9ead-5848a5ffdf24/16785811-1927-408b-9ead-5848a5ffdf24.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.838581] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d10923d7-ec50-4612-8779-82eac516196a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.845661] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1006.845661] env[62552]: value = "task-1239938" [ 1006.845661] env[62552]: _type = "Task" [ 1006.845661] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.855667] env[62552]: DEBUG nova.scheduler.client.report [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1006.859213] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239938, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.908023] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.944566] env[62552]: DEBUG oslo_vmware.api [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239935, 'name': ReconfigVM_Task, 'duration_secs': 0.152845} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.944899] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267504', 'volume_id': '418c407e-7b74-49e2-8014-086b46b72f2a', 'name': 'volume-418c407e-7b74-49e2-8014-086b46b72f2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '79166f2c-f864-4d8a-b1dc-e176710400cb', 'attached_at': '', 'detached_at': '', 'volume_id': '418c407e-7b74-49e2-8014-086b46b72f2a', 'serial': '418c407e-7b74-49e2-8014-086b46b72f2a'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1006.964557] env[62552]: DEBUG nova.network.neutron [req-0cb27016-1b07-4312-ac6c-7cff24f13558 req-fb665255-8743-44fe-9a43-303862bf370d service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Updated VIF entry in instance network info cache for port 9cf23ecf-7022-40a4-9182-2ac0d24f6fe4. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1006.964964] env[62552]: DEBUG nova.network.neutron [req-0cb27016-1b07-4312-ac6c-7cff24f13558 req-fb665255-8743-44fe-9a43-303862bf370d service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Updating instance_info_cache with network_info: [{"id": "9cf23ecf-7022-40a4-9182-2ac0d24f6fe4", "address": "fa:16:3e:3a:8b:84", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cf23ecf-70", "ovs_interfaceid": "9cf23ecf-7022-40a4-9182-2ac0d24f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.066078] env[62552]: INFO nova.compute.manager [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Took 15.57 seconds to build instance. [ 1007.286313] env[62552]: DEBUG nova.compute.manager [req-27a079b2-e80a-46e6-9cb8-60b3ce6836ac req-304c8f18-52e5-4430-85e5-68cd2cc55cee service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Received event network-vif-plugged-85efda2a-fdf8-4118-861c-af85ee3f0d46 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1007.286313] env[62552]: DEBUG oslo_concurrency.lockutils [req-27a079b2-e80a-46e6-9cb8-60b3ce6836ac req-304c8f18-52e5-4430-85e5-68cd2cc55cee service nova] Acquiring lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.286313] env[62552]: DEBUG oslo_concurrency.lockutils [req-27a079b2-e80a-46e6-9cb8-60b3ce6836ac req-304c8f18-52e5-4430-85e5-68cd2cc55cee service nova] Lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.287085] env[62552]: DEBUG oslo_concurrency.lockutils [req-27a079b2-e80a-46e6-9cb8-60b3ce6836ac req-304c8f18-52e5-4430-85e5-68cd2cc55cee service nova] Lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.287708] env[62552]: DEBUG nova.compute.manager [req-27a079b2-e80a-46e6-9cb8-60b3ce6836ac req-304c8f18-52e5-4430-85e5-68cd2cc55cee service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] No waiting events found dispatching network-vif-plugged-85efda2a-fdf8-4118-861c-af85ee3f0d46 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1007.288126] env[62552]: WARNING nova.compute.manager [req-27a079b2-e80a-46e6-9cb8-60b3ce6836ac req-304c8f18-52e5-4430-85e5-68cd2cc55cee service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Received unexpected event network-vif-plugged-85efda2a-fdf8-4118-861c-af85ee3f0d46 for instance with vm_state building and task_state spawning. [ 1007.331749] env[62552]: DEBUG oslo_vmware.api [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239937, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136595} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.333132] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1007.333507] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1007.333979] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1007.334321] env[62552]: INFO nova.compute.manager [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1007.334709] env[62552]: DEBUG oslo.service.loopingcall [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.336018] env[62552]: DEBUG nova.compute.manager [-] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1007.336018] env[62552]: DEBUG nova.network.neutron [-] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1007.362182] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.365180] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239938, 'name': Rename_Task, 'duration_secs': 0.159938} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.365693] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1007.365924] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c66b86a-4a54-4ef3-8231-213c4645e162 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.372136] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1007.372136] env[62552]: value = "task-1239939" [ 1007.372136] env[62552]: _type = "Task" [ 1007.372136] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.380126] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.389784] env[62552]: INFO nova.scheduler.client.report [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted allocations for instance 505d7f8e-c27f-487e-98d8-c840a526d40e [ 1007.397989] env[62552]: DEBUG nova.network.neutron [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Successfully updated port: 85efda2a-fdf8-4118-861c-af85ee3f0d46 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.421445] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.468023] env[62552]: DEBUG oslo_concurrency.lockutils [req-0cb27016-1b07-4312-ac6c-7cff24f13558 req-fb665255-8743-44fe-9a43-303862bf370d service nova] Releasing lock "refresh_cache-38bf968c-e757-4b46-b069-f625f5041c84" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.567440] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ecfbd827-b088-4d17-94f5-f842684b98e3 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.761s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.883631] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239939, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.898791] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e978b32e-bee1-4f3a-a818-4716b9f791e0 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "505d7f8e-c27f-487e-98d8-c840a526d40e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.696s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.914530] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.914530] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.914530] env[62552]: DEBUG nova.network.neutron [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1007.914530] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.993387] env[62552]: DEBUG nova.objects.instance [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lazy-loading 'flavor' on Instance uuid 79166f2c-f864-4d8a-b1dc-e176710400cb {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.189910] env[62552]: DEBUG nova.network.neutron [-] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.382173] env[62552]: DEBUG oslo_vmware.api [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239939, 'name': PowerOnVM_Task, 'duration_secs': 0.789037} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.382537] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1008.382735] env[62552]: INFO nova.compute.manager [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Took 8.13 seconds to spawn the instance on the hypervisor. [ 1008.382991] env[62552]: DEBUG nova.compute.manager [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1008.383820] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b011cef-480b-42af-b2eb-70bc812a7c50 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.414425] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.491994] env[62552]: DEBUG nova.network.neutron [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.506210] env[62552]: DEBUG oslo_concurrency.lockutils [None req-94a1952e-3883-4fca-9dbf-47d1fa70a2a1 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.380s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.589853] env[62552]: INFO nova.compute.manager [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Rescuing [ 1008.590174] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.590335] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.590893] env[62552]: DEBUG nova.network.neutron [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.692956] env[62552]: INFO nova.compute.manager [-] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Took 1.36 seconds to deallocate network for instance. [ 1008.909553] env[62552]: INFO nova.compute.manager [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Took 16.37 seconds to build instance. [ 1008.918542] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.929694] env[62552]: DEBUG nova.network.neutron [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Updating instance_info_cache with network_info: [{"id": "85efda2a-fdf8-4118-861c-af85ee3f0d46", "address": "fa:16:3e:76:c0:fd", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85efda2a-fd", "ovs_interfaceid": "85efda2a-fdf8-4118-861c-af85ee3f0d46", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.205070] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.205320] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.205597] env[62552]: DEBUG nova.objects.instance [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lazy-loading 'resources' on Instance uuid c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.402829] env[62552]: DEBUG nova.compute.manager [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Received event network-changed-85efda2a-fdf8-4118-861c-af85ee3f0d46 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1009.403558] env[62552]: DEBUG nova.compute.manager [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Refreshing instance network info cache due to event network-changed-85efda2a-fdf8-4118-861c-af85ee3f0d46. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1009.403806] env[62552]: DEBUG oslo_concurrency.lockutils [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] Acquiring lock "refresh_cache-2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.415082] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9c410f47-7ca8-408c-9741-bdf9b1c7cd0f tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "16785811-1927-408b-9ead-5848a5ffdf24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.892s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.415344] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.434095] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.434095] env[62552]: DEBUG nova.compute.manager [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Instance network_info: |[{"id": "85efda2a-fdf8-4118-861c-af85ee3f0d46", "address": "fa:16:3e:76:c0:fd", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85efda2a-fd", "ovs_interfaceid": "85efda2a-fdf8-4118-861c-af85ee3f0d46", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1009.434095] env[62552]: DEBUG oslo_concurrency.lockutils [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] Acquired lock "refresh_cache-2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.434095] env[62552]: DEBUG nova.network.neutron [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Refreshing network info cache for port 85efda2a-fdf8-4118-861c-af85ee3f0d46 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.435245] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:c0:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '85efda2a-fdf8-4118-861c-af85ee3f0d46', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.443129] env[62552]: DEBUG oslo.service.loopingcall [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.446756] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1009.447027] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0578bc26-98f9-4c13-9a2f-793042563a98 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.471563] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.471563] env[62552]: value = "task-1239940" [ 1009.471563] env[62552]: _type = "Task" [ 1009.471563] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.479181] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239940, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.501017] env[62552]: DEBUG nova.network.neutron [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating instance_info_cache with network_info: [{"id": "19a87259-1f7a-428c-8790-b74e27fde6b6", "address": "fa:16:3e:04:c8:af", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19a87259-1f", "ovs_interfaceid": "19a87259-1f7a-428c-8790-b74e27fde6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.745114] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "462a3224-d843-45f1-a2bf-69afc18e2572" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.745429] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.913332] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.952754] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dadf253-ab57-4600-892f-38708368a148 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.959696] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139b0be7-fe1c-4085-9549-a67477d1bac4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.994832] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabc4375-944b-4935-b8db-3be2c4717548 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.002685] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239940, 'name': CreateVM_Task, 'duration_secs': 0.446733} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.004595] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1010.005077] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.007366] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.007537] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.007844] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.009046] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbbbe78-1595-43ae-80a8-0e93414805bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.013275] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa166f91-7a48-4f5c-8e0a-de540809e8e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.017505] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1010.017505] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5222ab71-784f-bd20-8861-11016b0e7986" [ 1010.017505] env[62552]: _type = "Task" [ 1010.017505] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.025143] env[62552]: DEBUG nova.compute.provider_tree [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.035998] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5222ab71-784f-bd20-8861-11016b0e7986, 'name': SearchDatastore_Task, 'duration_secs': 0.010139} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.036879] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.037131] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.037374] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.037563] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.037900] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.040941] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c966f75-6b3c-487c-bbc7-8427737af495 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.049996] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.050213] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1010.051496] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d19fa768-ebdd-4235-a0a2-69e1b709a366 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.058922] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1010.058922] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529deee0-2ecc-8f35-1267-7e2cf138c118" [ 1010.058922] env[62552]: _type = "Task" [ 1010.058922] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.066624] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529deee0-2ecc-8f35-1267-7e2cf138c118, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.248457] env[62552]: DEBUG nova.compute.manager [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1010.272036] env[62552]: DEBUG nova.network.neutron [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Updated VIF entry in instance network info cache for port 85efda2a-fdf8-4118-861c-af85ee3f0d46. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.272444] env[62552]: DEBUG nova.network.neutron [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Updating instance_info_cache with network_info: [{"id": "85efda2a-fdf8-4118-861c-af85ee3f0d46", "address": "fa:16:3e:76:c0:fd", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85efda2a-fd", "ovs_interfaceid": "85efda2a-fdf8-4118-861c-af85ee3f0d46", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.365825] env[62552]: DEBUG nova.compute.manager [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Stashing vm_state: active {{(pid=62552) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 1010.414808] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.532595] env[62552]: DEBUG nova.scheduler.client.report [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1010.568331] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529deee0-2ecc-8f35-1267-7e2cf138c118, 'name': SearchDatastore_Task, 'duration_secs': 0.008816} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.569117] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee4582af-abc2-4241-b86e-45cef267d0be {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.574189] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1010.574189] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5213d911-78d7-8378-6278-c2b45b41347a" [ 1010.574189] env[62552]: _type = "Task" [ 1010.574189] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.581466] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5213d911-78d7-8378-6278-c2b45b41347a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.618485] env[62552]: DEBUG oslo_concurrency.lockutils [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "16785811-1927-408b-9ead-5848a5ffdf24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.618735] env[62552]: DEBUG oslo_concurrency.lockutils [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "16785811-1927-408b-9ead-5848a5ffdf24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.618945] env[62552]: DEBUG oslo_concurrency.lockutils [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "16785811-1927-408b-9ead-5848a5ffdf24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.619154] env[62552]: DEBUG oslo_concurrency.lockutils [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "16785811-1927-408b-9ead-5848a5ffdf24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.619328] env[62552]: DEBUG oslo_concurrency.lockutils [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "16785811-1927-408b-9ead-5848a5ffdf24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.621988] env[62552]: INFO nova.compute.manager [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Terminating instance [ 1010.766556] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.775285] env[62552]: DEBUG oslo_concurrency.lockutils [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] Releasing lock "refresh_cache-2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.775566] env[62552]: DEBUG nova.compute.manager [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Received event network-vif-deleted-e3f1462f-f27a-413e-b321-ec94895ff702 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1010.775772] env[62552]: DEBUG nova.compute.manager [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received event network-changed-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1010.775944] env[62552]: DEBUG nova.compute.manager [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Refreshing instance network info cache due to event network-changed-dd1028f0-d5a9-48be-84e3-9401cde0a1ff. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1010.776195] env[62552]: DEBUG oslo_concurrency.lockutils [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] Acquiring lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.776352] env[62552]: DEBUG oslo_concurrency.lockutils [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] Acquired lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.776528] env[62552]: DEBUG nova.network.neutron [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Refreshing network info cache for port dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1010.883875] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.916982] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.037455] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.039751] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.273s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.041323] env[62552]: INFO nova.compute.claims [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.049958] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.050240] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-92e36b5e-5d8b-441b-a0c8-37fe287e2302 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.057911] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1011.057911] env[62552]: value = "task-1239941" [ 1011.057911] env[62552]: _type = "Task" [ 1011.057911] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.058767] env[62552]: INFO nova.scheduler.client.report [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Deleted allocations for instance c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e [ 1011.069951] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.084592] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5213d911-78d7-8378-6278-c2b45b41347a, 'name': SearchDatastore_Task, 'duration_secs': 0.008811} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.084863] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.085150] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5/2f30cf85-df40-4560-8cc3-d0cdf7cbecc5.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1011.085416] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31964454-d421-45b6-a4b2-236f2fd7dcaf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.092430] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1011.092430] env[62552]: value = "task-1239942" [ 1011.092430] env[62552]: _type = "Task" [ 1011.092430] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.100629] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239942, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.125875] env[62552]: DEBUG nova.compute.manager [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1011.126114] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1011.127153] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d9ffab-116f-443e-bcac-086309406301 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.135012] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.135307] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2309c339-c749-45d0-9c9f-0f0290b64897 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.141650] env[62552]: DEBUG oslo_vmware.api [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1011.141650] env[62552]: value = "task-1239943" [ 1011.141650] env[62552]: _type = "Task" [ 1011.141650] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.149649] env[62552]: DEBUG oslo_vmware.api [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239943, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.423069] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.572161] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6e1f3cc9-22fb-433b-b6e3-f02726384e31 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.889s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.586042] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239941, 'name': PowerOffVM_Task, 'duration_secs': 0.190961} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.589583] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.590570] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68242eac-97f9-4f7a-8f6c-7a4a2ebfb25f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.627036] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd56ac4-eab5-4aea-8292-430ca6d67dc4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.630033] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239942, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.652805] env[62552]: DEBUG oslo_vmware.api [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239943, 'name': PowerOffVM_Task, 'duration_secs': 0.220494} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.655437] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.655678] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1011.657865] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d993ecd1-a1a2-467b-98d5-ad653e77f2d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.667062] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.667389] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b484bd57-29ab-45f3-bce1-75ef34ca60e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.673656] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1011.673656] env[62552]: value = "task-1239945" [ 1011.673656] env[62552]: _type = "Task" [ 1011.673656] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.685968] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1011.686607] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1011.686607] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.686717] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.687332] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1011.687332] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cc08b658-98ed-4da0-865d-1f8c2c712caa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.697403] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1011.698039] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1011.698608] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56554146-35b5-4baf-8374-c3ece68dfc8e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.704949] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1011.704949] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ca4f4f-fabe-8737-80a4-b306223285a9" [ 1011.704949] env[62552]: _type = "Task" [ 1011.704949] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.705799] env[62552]: DEBUG nova.network.neutron [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updated VIF entry in instance network info cache for port dd1028f0-d5a9-48be-84e3-9401cde0a1ff. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1011.706281] env[62552]: DEBUG nova.network.neutron [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updating instance_info_cache with network_info: [{"id": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "address": "fa:16:3e:38:6f:9c", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1028f0-d5", "ovs_interfaceid": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.716034] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ca4f4f-fabe-8737-80a4-b306223285a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.744401] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1011.744637] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1011.744847] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleting the datastore file [datastore2] 16785811-1927-408b-9ead-5848a5ffdf24 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.745126] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4330540-80ee-4fe5-b01b-27e10db356ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.751783] env[62552]: DEBUG oslo_vmware.api [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for the task: (returnval){ [ 1011.751783] env[62552]: value = "task-1239946" [ 1011.751783] env[62552]: _type = "Task" [ 1011.751783] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.760138] env[62552]: DEBUG oslo_vmware.api [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.917392] env[62552]: DEBUG oslo_vmware.api [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239932, 'name': ReconfigVM_Task, 'duration_secs': 5.802281} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.917739] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.921147] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Reconfigured VM to detach interface {{(pid=62552) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1012.107263] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239942, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531376} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.110010] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5/2f30cf85-df40-4560-8cc3-d0cdf7cbecc5.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1012.110341] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.110877] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09c16dd2-ec2e-488c-89ef-6bca151bbd0d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.117793] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1012.117793] env[62552]: value = "task-1239947" [ 1012.117793] env[62552]: _type = "Task" [ 1012.117793] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.131230] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239947, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.212801] env[62552]: DEBUG oslo_concurrency.lockutils [req-6b8a5af5-15ec-4c5c-8cfd-3e69250fc200 req-93c7828b-0218-41d2-a4dd-ec859acd8c4b service nova] Releasing lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.223929] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52ca4f4f-fabe-8737-80a4-b306223285a9, 'name': SearchDatastore_Task, 'duration_secs': 0.010215} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.224493] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4295b50b-067e-409c-9c66-aadb65ee8b04 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.233537] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1012.233537] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52966819-1212-e8eb-750e-d952f9df25fc" [ 1012.233537] env[62552]: _type = "Task" [ 1012.233537] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.243365] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52966819-1212-e8eb-750e-d952f9df25fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.266590] env[62552]: DEBUG oslo_vmware.api [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Task: {'id': task-1239946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151824} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.266863] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.267072] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1012.267262] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1012.267449] env[62552]: INFO nova.compute.manager [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1012.267679] env[62552]: DEBUG oslo.service.loopingcall [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.267873] env[62552]: DEBUG nova.compute.manager [-] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1012.267969] env[62552]: DEBUG nova.network.neutron [-] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1012.287283] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709508be-553a-4e00-aac3-933adb8e1e2c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.302550] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467c076d-e163-4ee1-bc4e-57743908104f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.334507] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff64cc5-4a66-4efb-b7de-a644cb4d4cd6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.341895] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac8fff7-5f84-45ca-95b9-83a29d76c93b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.355887] env[62552]: DEBUG nova.compute.provider_tree [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.629745] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239947, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109767} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.629745] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.629745] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1278bf4-6ddd-4587-9dfd-75ee61f093f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.653932] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5/2f30cf85-df40-4560-8cc3-d0cdf7cbecc5.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.654216] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45c181af-f999-4a55-acd3-6d61b1bcd197 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.675985] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1012.675985] env[62552]: value = "task-1239948" [ 1012.675985] env[62552]: _type = "Task" [ 1012.675985] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.684392] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239948, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.745220] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52966819-1212-e8eb-750e-d952f9df25fc, 'name': SearchDatastore_Task, 'duration_secs': 0.032459} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.745698] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.745779] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 79166f2c-f864-4d8a-b1dc-e176710400cb/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. {{(pid=62552) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1012.746072] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9880ac03-872a-4dc1-aaaf-c2edad096738 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.752656] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1012.752656] env[62552]: value = "task-1239949" [ 1012.752656] env[62552]: _type = "Task" [ 1012.752656] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.761217] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239949, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.860151] env[62552]: DEBUG nova.scheduler.client.report [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1012.872296] env[62552]: DEBUG nova.compute.manager [req-30d4df79-bae8-421e-b3a8-cd7953478cbc req-7cb6d54e-ec6a-413c-8173-5602a9f1e356 service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Received event network-vif-deleted-d05003ff-4c67-4508-a70a-debf5ecec19e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1012.872525] env[62552]: INFO nova.compute.manager [req-30d4df79-bae8-421e-b3a8-cd7953478cbc req-7cb6d54e-ec6a-413c-8173-5602a9f1e356 service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Neutron deleted interface d05003ff-4c67-4508-a70a-debf5ecec19e; detaching it from the instance and deleting it from the info cache [ 1012.872712] env[62552]: DEBUG nova.network.neutron [req-30d4df79-bae8-421e-b3a8-cd7953478cbc req-7cb6d54e-ec6a-413c-8173-5602a9f1e356 service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.186965] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239948, 'name': ReconfigVM_Task, 'duration_secs': 0.420407} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.187306] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5/2f30cf85-df40-4560-8cc3-d0cdf7cbecc5.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.187968] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b6a8230-1fea-406f-81c9-e4dcdc24c080 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.195316] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1013.195316] env[62552]: value = "task-1239950" [ 1013.195316] env[62552]: _type = "Task" [ 1013.195316] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.203750] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239950, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.262060] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239949, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470854} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.262416] env[62552]: INFO nova.virt.vmwareapi.ds_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 79166f2c-f864-4d8a-b1dc-e176710400cb/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. [ 1013.263204] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1044469c-efa5-454a-9704-16d34b2dbc84 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.290730] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 79166f2c-f864-4d8a-b1dc-e176710400cb/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.291066] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b0ad38e-4e8e-478c-878e-084680c2edca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.304026] env[62552]: DEBUG nova.network.neutron [-] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.309794] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1013.309794] env[62552]: value = "task-1239951" [ 1013.309794] env[62552]: _type = "Task" [ 1013.309794] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.318983] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.369179] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.369855] env[62552]: DEBUG nova.compute.manager [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1013.373965] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.490s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.380243] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37ce4953-1ac2-4c14-b38b-e41e9fd67df9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.383072] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "19b69199-b3e1-45c5-af35-d9329113d360" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.383343] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "19b69199-b3e1-45c5-af35-d9329113d360" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.394476] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15515817-c407-4e20-afd0-8a1f974248ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.428568] env[62552]: DEBUG nova.compute.manager [req-30d4df79-bae8-421e-b3a8-cd7953478cbc req-7cb6d54e-ec6a-413c-8173-5602a9f1e356 service nova] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Detach interface failed, port_id=d05003ff-4c67-4508-a70a-debf5ecec19e, reason: Instance 16785811-1927-408b-9ead-5848a5ffdf24 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1013.442022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.442022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.442022] env[62552]: DEBUG nova.network.neutron [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1013.704857] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239950, 'name': Rename_Task, 'duration_secs': 0.229966} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.705213] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1013.705460] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-791808a8-ad2d-4c39-8c82-8cc10fe0c583 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.711419] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1013.711419] env[62552]: value = "task-1239952" [ 1013.711419] env[62552]: _type = "Task" [ 1013.711419] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.718853] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239952, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.806605] env[62552]: INFO nova.compute.manager [-] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Took 1.54 seconds to deallocate network for instance. [ 1013.820879] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239951, 'name': ReconfigVM_Task, 'duration_secs': 0.349262} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.821189] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 79166f2c-f864-4d8a-b1dc-e176710400cb/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.822438] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7644da8-6426-4078-9f32-366c5fcc5811 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.852515] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72274098-eb7b-43e9-9b14-4247d8d61430 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.868860] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1013.868860] env[62552]: value = "task-1239953" [ 1013.868860] env[62552]: _type = "Task" [ 1013.868860] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.878268] env[62552]: DEBUG nova.compute.utils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.879615] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239953, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.880243] env[62552]: DEBUG nova.compute.manager [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1013.880411] env[62552]: DEBUG nova.network.neutron [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1013.884489] env[62552]: INFO nova.compute.claims [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.888372] env[62552]: DEBUG nova.compute.manager [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1014.037727] env[62552]: DEBUG nova.policy [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb432303c2394f8795672ceaf02f28aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a0bc011d6794602b2bbe1fc01e4c8b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1014.229988] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239952, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.316490] env[62552]: DEBUG oslo_concurrency.lockutils [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.379765] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239953, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.381380] env[62552]: DEBUG nova.compute.manager [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1014.395654] env[62552]: INFO nova.compute.resource_tracker [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating resource usage from migration 67354786-4ad2-4dc7-8bd8-ceee01cc04a1 [ 1014.425101] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.654343] env[62552]: INFO nova.network.neutron [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Port 3688687a-8e5c-4347-8972-724ebb0df317 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1014.654797] env[62552]: DEBUG nova.network.neutron [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.662574] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9191d517-9e04-42b1-ac73-8dc6261b727d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.671827] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d662253d-03d0-47a9-b232-3450c4ec53ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.709203] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2b69ea-e0a1-4064-90d3-78e1fd010309 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.720017] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2449b32-5609-417d-96d8-e818ea3ff0db {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.729272] env[62552]: DEBUG oslo_vmware.api [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239952, 'name': PowerOnVM_Task, 'duration_secs': 0.721453} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.738137] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1014.738643] env[62552]: INFO nova.compute.manager [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Took 8.19 seconds to spawn the instance on the hypervisor. [ 1014.738901] env[62552]: DEBUG nova.compute.manager [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1014.739856] env[62552]: DEBUG nova.compute.provider_tree [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.741826] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080dced4-e712-4637-ab93-125823ab25de {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.883509] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239953, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.979615] env[62552]: DEBUG nova.network.neutron [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Successfully created port: 81435491-9993-42c8-80d8-8c4f23693ce4 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.099398] env[62552]: DEBUG oslo_concurrency.lockutils [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquiring lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.099665] env[62552]: DEBUG oslo_concurrency.lockutils [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.099911] env[62552]: DEBUG oslo_concurrency.lockutils [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquiring lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.101083] env[62552]: DEBUG oslo_concurrency.lockutils [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.101083] env[62552]: DEBUG oslo_concurrency.lockutils [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.103096] env[62552]: INFO nova.compute.manager [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Terminating instance [ 1015.157596] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.167974] env[62552]: DEBUG nova.compute.manager [req-c6b27221-000c-4c4f-80ec-7f773a7fd7b2 req-18b9af36-0931-4801-ab91-87c8b36dc853 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1015.168199] env[62552]: DEBUG nova.compute.manager [req-c6b27221-000c-4c4f-80ec-7f773a7fd7b2 req-18b9af36-0931-4801-ab91-87c8b36dc853 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing instance network info cache due to event network-changed-bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1015.168792] env[62552]: DEBUG oslo_concurrency.lockutils [req-c6b27221-000c-4c4f-80ec-7f773a7fd7b2 req-18b9af36-0931-4801-ab91-87c8b36dc853 service nova] Acquiring lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.168976] env[62552]: DEBUG oslo_concurrency.lockutils [req-c6b27221-000c-4c4f-80ec-7f773a7fd7b2 req-18b9af36-0931-4801-ab91-87c8b36dc853 service nova] Acquired lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.169439] env[62552]: DEBUG nova.network.neutron [req-c6b27221-000c-4c4f-80ec-7f773a7fd7b2 req-18b9af36-0931-4801-ab91-87c8b36dc853 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Refreshing network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.246705] env[62552]: DEBUG nova.scheduler.client.report [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1015.265278] env[62552]: INFO nova.compute.manager [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Took 19.10 seconds to build instance. [ 1015.388484] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239953, 'name': ReconfigVM_Task, 'duration_secs': 1.085597} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.388803] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1015.389111] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd2f14be-d328-40d1-b2cf-d783e0501ded {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.393551] env[62552]: DEBUG nova.compute.manager [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1015.398140] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1015.398140] env[62552]: value = "task-1239954" [ 1015.398140] env[62552]: _type = "Task" [ 1015.398140] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.410058] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.431774] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.432053] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.432270] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.432488] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.432645] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.432798] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.433041] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.433203] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.433376] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.433544] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.433743] env[62552]: DEBUG nova.virt.hardware [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.435019] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1c4763-7fe5-4039-b9af-eadb615e0c94 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.442519] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e18712-37d7-4d5e-8a0e-60da19e404a0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.607183] env[62552]: DEBUG nova.compute.manager [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1015.607452] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.608377] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ed17b1-0dc0-4e99-944f-c9e5c121a8d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.617212] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.617212] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6569ea8-061a-4864-85cc-2e3deaa9f90a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.623253] env[62552]: DEBUG oslo_vmware.api [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 1015.623253] env[62552]: value = "task-1239955" [ 1015.623253] env[62552]: _type = "Task" [ 1015.623253] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.631578] env[62552]: DEBUG oslo_vmware.api [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.661971] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b789cc2-c20f-4240-9ca7-0d80fb0d6278 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-6a4573f0-7152-4462-bb72-58a45b2cbd97-3688687a-8e5c-4347-8972-724ebb0df317" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.461s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.755757] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.382s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.756122] env[62552]: INFO nova.compute.manager [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Migrating [ 1015.764497] env[62552]: DEBUG oslo_concurrency.lockutils [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.448s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.765170] env[62552]: DEBUG nova.objects.instance [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lazy-loading 'resources' on Instance uuid 16785811-1927-408b-9ead-5848a5ffdf24 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.768758] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a02e77a0-a8c8-452a-9585-bec554bf0f0b tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.611s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.908385] env[62552]: DEBUG oslo_vmware.api [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239954, 'name': PowerOnVM_Task, 'duration_secs': 0.424515} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.911149] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.914211] env[62552]: DEBUG nova.compute.manager [None req-f70c5e92-c620-4f92-be4b-1d44d92f2d66 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1015.915021] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80512324-4029-41e9-9824-c4e53936f2d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.995068] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "interface-d0aa144e-9bcf-4faa-8d4c-7743118fbe69-3688687a-8e5c-4347-8972-724ebb0df317" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.995250] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-d0aa144e-9bcf-4faa-8d4c-7743118fbe69-3688687a-8e5c-4347-8972-724ebb0df317" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.995605] env[62552]: DEBUG nova.objects.instance [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'flavor' on Instance uuid d0aa144e-9bcf-4faa-8d4c-7743118fbe69 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.025470] env[62552]: DEBUG nova.network.neutron [req-c6b27221-000c-4c4f-80ec-7f773a7fd7b2 req-18b9af36-0931-4801-ab91-87c8b36dc853 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updated VIF entry in instance network info cache for port bb66aa2b-bd30-454a-b71b-da2a0285cef1. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1016.025847] env[62552]: DEBUG nova.network.neutron [req-c6b27221-000c-4c4f-80ec-7f773a7fd7b2 req-18b9af36-0931-4801-ab91-87c8b36dc853 service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [{"id": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "address": "fa:16:3e:7d:2a:ae", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb66aa2b-bd", "ovs_interfaceid": "bb66aa2b-bd30-454a-b71b-da2a0285cef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.054018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b692c5f8-7bb8-4f94-9bb2-4ed72a7989ed {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.058712] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6e39d636-0dd1-43d8-abcc-bee4a33fa174 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Suspending the VM {{(pid=62552) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1016.059057] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-b1e5e505-a27e-4be4-a753-e18805c8b626 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.065321] env[62552]: DEBUG oslo_vmware.api [None req-6e39d636-0dd1-43d8-abcc-bee4a33fa174 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1016.065321] env[62552]: value = "task-1239956" [ 1016.065321] env[62552]: _type = "Task" [ 1016.065321] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.075712] env[62552]: DEBUG oslo_vmware.api [None req-6e39d636-0dd1-43d8-abcc-bee4a33fa174 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239956, 'name': SuspendVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.132911] env[62552]: DEBUG oslo_vmware.api [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239955, 'name': PowerOffVM_Task, 'duration_secs': 0.208803} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.133399] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1016.133714] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1016.134096] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7270e296-dbdc-41b6-893f-545ec55612e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.278053] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.278471] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.278754] env[62552]: DEBUG nova.network.neutron [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.286021] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1016.286367] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1016.286615] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Deleting the datastore file [datastore2] 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1016.286933] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a845ec7f-1d70-497d-83f5-47dc31a2c558 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.301161] env[62552]: DEBUG oslo_vmware.api [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for the task: (returnval){ [ 1016.301161] env[62552]: value = "task-1239958" [ 1016.301161] env[62552]: _type = "Task" [ 1016.301161] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.310362] env[62552]: DEBUG oslo_vmware.api [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.513163] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dcadd38-1990-49f5-a0d7-b9f394e01794 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.521057] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4f066e-18cd-493d-beea-34b0e7d71d4b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.550045] env[62552]: DEBUG oslo_concurrency.lockutils [req-c6b27221-000c-4c4f-80ec-7f773a7fd7b2 req-18b9af36-0931-4801-ab91-87c8b36dc853 service nova] Releasing lock "refresh_cache-6a4573f0-7152-4462-bb72-58a45b2cbd97" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.551089] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784bde8d-445f-4ddb-8430-973a65e4924e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.560326] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e508c7-a910-4363-9cdb-45db26f90e70 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.574120] env[62552]: DEBUG nova.compute.provider_tree [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.583429] env[62552]: DEBUG oslo_vmware.api [None req-6e39d636-0dd1-43d8-abcc-bee4a33fa174 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239956, 'name': SuspendVM_Task} progress is 54%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.711251] env[62552]: DEBUG nova.objects.instance [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'pci_requests' on Instance uuid d0aa144e-9bcf-4faa-8d4c-7743118fbe69 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.812474] env[62552]: DEBUG oslo_vmware.api [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Task: {'id': task-1239958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18471} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.812858] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.813218] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.813519] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.813802] env[62552]: INFO nova.compute.manager [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1016.814159] env[62552]: DEBUG oslo.service.loopingcall [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.814458] env[62552]: DEBUG nova.compute.manager [-] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1016.814663] env[62552]: DEBUG nova.network.neutron [-] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1016.864321] env[62552]: DEBUG nova.compute.manager [req-153c6ffd-a36f-4c37-aa1d-011a567f0801 req-da0b2f66-f811-4b79-b76c-73a2a845cd82 service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Received event network-vif-plugged-81435491-9993-42c8-80d8-8c4f23693ce4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1016.864548] env[62552]: DEBUG oslo_concurrency.lockutils [req-153c6ffd-a36f-4c37-aa1d-011a567f0801 req-da0b2f66-f811-4b79-b76c-73a2a845cd82 service nova] Acquiring lock "462a3224-d843-45f1-a2bf-69afc18e2572-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.865000] env[62552]: DEBUG oslo_concurrency.lockutils [req-153c6ffd-a36f-4c37-aa1d-011a567f0801 req-da0b2f66-f811-4b79-b76c-73a2a845cd82 service nova] Lock "462a3224-d843-45f1-a2bf-69afc18e2572-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.865336] env[62552]: DEBUG oslo_concurrency.lockutils [req-153c6ffd-a36f-4c37-aa1d-011a567f0801 req-da0b2f66-f811-4b79-b76c-73a2a845cd82 service nova] Lock "462a3224-d843-45f1-a2bf-69afc18e2572-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.865647] env[62552]: DEBUG nova.compute.manager [req-153c6ffd-a36f-4c37-aa1d-011a567f0801 req-da0b2f66-f811-4b79-b76c-73a2a845cd82 service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] No waiting events found dispatching network-vif-plugged-81435491-9993-42c8-80d8-8c4f23693ce4 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1016.865927] env[62552]: WARNING nova.compute.manager [req-153c6ffd-a36f-4c37-aa1d-011a567f0801 req-da0b2f66-f811-4b79-b76c-73a2a845cd82 service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Received unexpected event network-vif-plugged-81435491-9993-42c8-80d8-8c4f23693ce4 for instance with vm_state building and task_state spawning. [ 1017.079995] env[62552]: DEBUG nova.scheduler.client.report [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1017.090739] env[62552]: DEBUG oslo_vmware.api [None req-6e39d636-0dd1-43d8-abcc-bee4a33fa174 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239956, 'name': SuspendVM_Task, 'duration_secs': 0.838392} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.094082] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6e39d636-0dd1-43d8-abcc-bee4a33fa174 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Suspended the VM {{(pid=62552) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1017.094082] env[62552]: DEBUG nova.compute.manager [None req-6e39d636-0dd1-43d8-abcc-bee4a33fa174 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1017.094082] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c745267-0d47-4bee-b196-d2535616f070 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.108462] env[62552]: DEBUG nova.network.neutron [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.211251] env[62552]: DEBUG nova.objects.base [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1017.211319] env[62552]: DEBUG nova.network.neutron [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1017.266509] env[62552]: DEBUG nova.compute.manager [req-a81c713f-7129-48b1-b5ab-1fd85fcfb17d req-e42c1325-2be9-445d-9306-715929c84983 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received event network-changed-34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1017.266717] env[62552]: DEBUG nova.compute.manager [req-a81c713f-7129-48b1-b5ab-1fd85fcfb17d req-e42c1325-2be9-445d-9306-715929c84983 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing instance network info cache due to event network-changed-34124ec0-5d30-4c38-8095-b11bda9f58fa. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1017.266922] env[62552]: DEBUG oslo_concurrency.lockutils [req-a81c713f-7129-48b1-b5ab-1fd85fcfb17d req-e42c1325-2be9-445d-9306-715929c84983 service nova] Acquiring lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.267103] env[62552]: DEBUG oslo_concurrency.lockutils [req-a81c713f-7129-48b1-b5ab-1fd85fcfb17d req-e42c1325-2be9-445d-9306-715929c84983 service nova] Acquired lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.267295] env[62552]: DEBUG nova.network.neutron [req-a81c713f-7129-48b1-b5ab-1fd85fcfb17d req-e42c1325-2be9-445d-9306-715929c84983 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing network info cache for port 34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.297541] env[62552]: DEBUG nova.compute.manager [req-a64b354c-0106-4636-a725-002e133a2b02 req-ab12d966-ec02-47f7-b5d8-f959b0bf9baf service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Received event network-vif-deleted-4294380b-fd37-48f3-ba91-6e169312e2a0 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1017.297716] env[62552]: INFO nova.compute.manager [req-a64b354c-0106-4636-a725-002e133a2b02 req-ab12d966-ec02-47f7-b5d8-f959b0bf9baf service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Neutron deleted interface 4294380b-fd37-48f3-ba91-6e169312e2a0; detaching it from the instance and deleting it from the info cache [ 1017.297900] env[62552]: DEBUG nova.network.neutron [req-a64b354c-0106-4636-a725-002e133a2b02 req-ab12d966-ec02-47f7-b5d8-f959b0bf9baf service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.300297] env[62552]: DEBUG nova.policy [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6515846d33e44bca942767f9e1d650b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2f57f72de0c4adc9754458f4e7040b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1017.364084] env[62552]: DEBUG nova.network.neutron [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Successfully updated port: 81435491-9993-42c8-80d8-8c4f23693ce4 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1017.585498] env[62552]: DEBUG oslo_concurrency.lockutils [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.821s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.587804] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.163s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.590489] env[62552]: INFO nova.compute.claims [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.612122] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.619954] env[62552]: INFO nova.scheduler.client.report [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Deleted allocations for instance 16785811-1927-408b-9ead-5848a5ffdf24 [ 1017.776025] env[62552]: DEBUG nova.network.neutron [-] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.804283] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e11a4a1-b092-4a2c-a2bf-fb0184bcde02 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.815213] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5b5c51-724d-406b-9399-502d3a86193a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.846938] env[62552]: DEBUG nova.compute.manager [req-a64b354c-0106-4636-a725-002e133a2b02 req-ab12d966-ec02-47f7-b5d8-f959b0bf9baf service nova] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Detach interface failed, port_id=4294380b-fd37-48f3-ba91-6e169312e2a0, reason: Instance 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1017.868473] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "refresh_cache-462a3224-d843-45f1-a2bf-69afc18e2572" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.868473] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "refresh_cache-462a3224-d843-45f1-a2bf-69afc18e2572" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.868473] env[62552]: DEBUG nova.network.neutron [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.986894] env[62552]: INFO nova.compute.manager [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Unrescuing [ 1017.986894] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.987083] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquired lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.987227] env[62552]: DEBUG nova.network.neutron [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1018.072403] env[62552]: DEBUG nova.network.neutron [req-a81c713f-7129-48b1-b5ab-1fd85fcfb17d req-e42c1325-2be9-445d-9306-715929c84983 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updated VIF entry in instance network info cache for port 34124ec0-5d30-4c38-8095-b11bda9f58fa. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1018.072941] env[62552]: DEBUG nova.network.neutron [req-a81c713f-7129-48b1-b5ab-1fd85fcfb17d req-e42c1325-2be9-445d-9306-715929c84983 service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [{"id": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "address": "fa:16:3e:c9:01:94", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34124ec0-5d", "ovs_interfaceid": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.127290] env[62552]: DEBUG oslo_concurrency.lockutils [None req-55966a75-9722-4042-893f-8ea4e234d49c tempest-ServerDiskConfigTestJSON-1917216167 tempest-ServerDiskConfigTestJSON-1917216167-project-member] Lock "16785811-1927-408b-9ead-5848a5ffdf24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.508s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.278229] env[62552]: INFO nova.compute.manager [-] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Took 1.46 seconds to deallocate network for instance. [ 1018.414771] env[62552]: DEBUG nova.network.neutron [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1018.544879] env[62552]: DEBUG nova.network.neutron [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Updating instance_info_cache with network_info: [{"id": "81435491-9993-42c8-80d8-8c4f23693ce4", "address": "fa:16:3e:c5:f9:cc", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81435491-99", "ovs_interfaceid": "81435491-9993-42c8-80d8-8c4f23693ce4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.578258] env[62552]: DEBUG oslo_concurrency.lockutils [req-a81c713f-7129-48b1-b5ab-1fd85fcfb17d req-e42c1325-2be9-445d-9306-715929c84983 service nova] Releasing lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.756941] env[62552]: DEBUG oslo_concurrency.lockutils [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.757260] env[62552]: DEBUG oslo_concurrency.lockutils [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.758149] env[62552]: DEBUG oslo_concurrency.lockutils [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.758149] env[62552]: DEBUG oslo_concurrency.lockutils [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.758149] env[62552]: DEBUG oslo_concurrency.lockutils [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.761085] env[62552]: INFO nova.compute.manager [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Terminating instance [ 1018.774062] env[62552]: DEBUG nova.network.neutron [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating instance_info_cache with network_info: [{"id": "19a87259-1f7a-428c-8790-b74e27fde6b6", "address": "fa:16:3e:04:c8:af", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19a87259-1f", "ovs_interfaceid": "19a87259-1f7a-428c-8790-b74e27fde6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.785523] env[62552]: DEBUG oslo_concurrency.lockutils [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.857202] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd6463b-1096-492b-a2c4-5f22d61c5015 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.867927] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aadc039-313c-47b7-ba14-2e0746577329 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.907461] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0252a64f-15fe-4223-ad68-17cae809e68a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.915628] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30191d7f-729e-469d-ab69-001aaa1cb920 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.933471] env[62552]: DEBUG nova.compute.provider_tree [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.048008] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "refresh_cache-462a3224-d843-45f1-a2bf-69afc18e2572" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.048422] env[62552]: DEBUG nova.compute.manager [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Instance network_info: |[{"id": "81435491-9993-42c8-80d8-8c4f23693ce4", "address": "fa:16:3e:c5:f9:cc", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81435491-99", "ovs_interfaceid": "81435491-9993-42c8-80d8-8c4f23693ce4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1019.050745] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:f9:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81435491-9993-42c8-80d8-8c4f23693ce4', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1019.060502] env[62552]: DEBUG oslo.service.loopingcall [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.060887] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1019.061175] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b41fbde-0c92-4730-8397-8919d3d7571e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.089599] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.089599] env[62552]: value = "task-1239959" [ 1019.089599] env[62552]: _type = "Task" [ 1019.089599] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.097891] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239959, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.131215] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24584c4-e0a9-4df9-9ab3-2bf424c84489 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.151334] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance '88d0f502-fc3f-429e-bdf6-a1ebacec2117' progress to 0 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1019.155590] env[62552]: DEBUG nova.network.neutron [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Successfully updated port: 3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1019.265515] env[62552]: DEBUG nova.compute.manager [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1019.265863] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1019.266807] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b5c10f-07c1-4196-816a-1da1a396414f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.274232] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1019.274502] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b241d67-8d72-4c0b-88f7-e2a0877b54a4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.276304] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Releasing lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.276918] env[62552]: DEBUG nova.objects.instance [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lazy-loading 'flavor' on Instance uuid 79166f2c-f864-4d8a-b1dc-e176710400cb {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.295775] env[62552]: DEBUG nova.compute.manager [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Received event network-changed-81435491-9993-42c8-80d8-8c4f23693ce4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1019.295968] env[62552]: DEBUG nova.compute.manager [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Refreshing instance network info cache due to event network-changed-81435491-9993-42c8-80d8-8c4f23693ce4. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1019.296215] env[62552]: DEBUG oslo_concurrency.lockutils [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] Acquiring lock "refresh_cache-462a3224-d843-45f1-a2bf-69afc18e2572" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.296362] env[62552]: DEBUG oslo_concurrency.lockutils [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] Acquired lock "refresh_cache-462a3224-d843-45f1-a2bf-69afc18e2572" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.296522] env[62552]: DEBUG nova.network.neutron [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Refreshing network info cache for port 81435491-9993-42c8-80d8-8c4f23693ce4 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1019.344122] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1019.344405] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1019.344633] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleting the datastore file [datastore1] 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.344907] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1c2479a-e36f-4c3c-a887-fd958d2e1a81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.351148] env[62552]: DEBUG oslo_vmware.api [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1019.351148] env[62552]: value = "task-1239961" [ 1019.351148] env[62552]: _type = "Task" [ 1019.351148] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.360081] env[62552]: DEBUG oslo_vmware.api [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239961, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.436888] env[62552]: DEBUG nova.scheduler.client.report [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1019.599209] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239959, 'name': CreateVM_Task, 'duration_secs': 0.378182} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.599385] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1019.600133] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.600322] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.600649] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1019.600916] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-883ecb13-c7dd-4fc6-9a2d-ea4a1cb2c4ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.605359] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1019.605359] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5249d734-a0d0-72e2-3962-af2cad98328b" [ 1019.605359] env[62552]: _type = "Task" [ 1019.605359] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.613187] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5249d734-a0d0-72e2-3962-af2cad98328b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.659196] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1019.659771] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.659931] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.660117] env[62552]: DEBUG nova.network.neutron [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1019.661337] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe82a12c-e5e0-4719-bfcd-329e1f566cda {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.668307] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1019.668307] env[62552]: value = "task-1239962" [ 1019.668307] env[62552]: _type = "Task" [ 1019.668307] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.677198] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239962, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.782875] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b3b3e2-9c61-4103-be79-6b80ae551376 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.815270] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1019.815878] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83749610-e2d4-4abe-b010-4c1de7b7d662 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.822579] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1019.822579] env[62552]: value = "task-1239963" [ 1019.822579] env[62552]: _type = "Task" [ 1019.822579] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.830583] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239963, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.860019] env[62552]: DEBUG oslo_vmware.api [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239961, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20014} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.860649] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.860649] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.860725] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1019.860861] env[62552]: INFO nova.compute.manager [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1019.861117] env[62552]: DEBUG oslo.service.loopingcall [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.861357] env[62552]: DEBUG nova.compute.manager [-] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1019.861472] env[62552]: DEBUG nova.network.neutron [-] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1019.941798] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.942471] env[62552]: DEBUG nova.compute.manager [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1019.947702] env[62552]: DEBUG oslo_concurrency.lockutils [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.162s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.948087] env[62552]: DEBUG nova.objects.instance [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lazy-loading 'resources' on Instance uuid 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.097242] env[62552]: DEBUG nova.network.neutron [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Updated VIF entry in instance network info cache for port 81435491-9993-42c8-80d8-8c4f23693ce4. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1020.097657] env[62552]: DEBUG nova.network.neutron [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Updating instance_info_cache with network_info: [{"id": "81435491-9993-42c8-80d8-8c4f23693ce4", "address": "fa:16:3e:c5:f9:cc", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81435491-99", "ovs_interfaceid": "81435491-9993-42c8-80d8-8c4f23693ce4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.116494] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5249d734-a0d0-72e2-3962-af2cad98328b, 'name': SearchDatastore_Task, 'duration_secs': 0.00958} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.116825] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.117089] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.117338] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.117486] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.117667] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.117952] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66f04e85-3973-4d3c-9662-2f3837faf2cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.125486] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.125669] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1020.126627] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-041a88f9-a4c3-4126-8663-2b2c7f6f6dec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.131650] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1020.131650] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5275303d-e97b-bb3c-d4ae-57bd8134e6ed" [ 1020.131650] env[62552]: _type = "Task" [ 1020.131650] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.138878] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5275303d-e97b-bb3c-d4ae-57bd8134e6ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.178274] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239962, 'name': PowerOffVM_Task, 'duration_secs': 0.420924} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.178646] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.178901] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance '88d0f502-fc3f-429e-bdf6-a1ebacec2117' progress to 17 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1020.206344] env[62552]: WARNING nova.network.neutron [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] 06f1ea1a-5103-419f-a356-3b65b9a2685e already exists in list: networks containing: ['06f1ea1a-5103-419f-a356-3b65b9a2685e']. ignoring it [ 1020.337243] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239963, 'name': PowerOffVM_Task, 'duration_secs': 0.271137} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.337641] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.344311] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfiguring VM instance instance-00000055 to detach disk 2002 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1020.344526] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6b49476-17da-4a31-9f79-0ca9a9ed0e75 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.363492] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1020.363492] env[62552]: value = "task-1239964" [ 1020.363492] env[62552]: _type = "Task" [ 1020.363492] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.372201] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239964, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.394196] env[62552]: DEBUG nova.compute.manager [req-16ab1deb-3b95-4013-bd6e-959effbe881e req-942cde9c-9398-48f2-9687-9fe18f56968d service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Received event network-vif-deleted-85efda2a-fdf8-4118-861c-af85ee3f0d46 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1020.394406] env[62552]: INFO nova.compute.manager [req-16ab1deb-3b95-4013-bd6e-959effbe881e req-942cde9c-9398-48f2-9687-9fe18f56968d service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Neutron deleted interface 85efda2a-fdf8-4118-861c-af85ee3f0d46; detaching it from the instance and deleting it from the info cache [ 1020.394575] env[62552]: DEBUG nova.network.neutron [req-16ab1deb-3b95-4013-bd6e-959effbe881e req-942cde9c-9398-48f2-9687-9fe18f56968d service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.453130] env[62552]: DEBUG nova.compute.utils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1020.454543] env[62552]: DEBUG nova.compute.manager [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1020.454776] env[62552]: DEBUG nova.network.neutron [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1020.527012] env[62552]: DEBUG nova.policy [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62819c95c0f14cad89513269b1f0bf1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa13c44376934f519998396c182a49f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1020.582974] env[62552]: DEBUG nova.network.neutron [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [{"id": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "address": "fa:16:3e:c9:01:94", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34124ec0-5d", "ovs_interfaceid": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3688687a-8e5c-4347-8972-724ebb0df317", "address": "fa:16:3e:79:0b:d3", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3688687a-8e", "ovs_interfaceid": "3688687a-8e5c-4347-8972-724ebb0df317", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.600922] env[62552]: DEBUG oslo_concurrency.lockutils [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] Releasing lock "refresh_cache-462a3224-d843-45f1-a2bf-69afc18e2572" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.601482] env[62552]: DEBUG nova.compute.manager [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received event network-vif-plugged-3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1020.601482] env[62552]: DEBUG oslo_concurrency.lockutils [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] Acquiring lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.601571] env[62552]: DEBUG oslo_concurrency.lockutils [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.601856] env[62552]: DEBUG oslo_concurrency.lockutils [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.601922] env[62552]: DEBUG nova.compute.manager [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] No waiting events found dispatching network-vif-plugged-3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1020.602387] env[62552]: WARNING nova.compute.manager [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received unexpected event network-vif-plugged-3688687a-8e5c-4347-8972-724ebb0df317 for instance with vm_state active and task_state None. [ 1020.602387] env[62552]: DEBUG nova.compute.manager [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received event network-changed-3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1020.602498] env[62552]: DEBUG nova.compute.manager [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing instance network info cache due to event network-changed-3688687a-8e5c-4347-8972-724ebb0df317. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1020.602652] env[62552]: DEBUG oslo_concurrency.lockutils [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] Acquiring lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.644063] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5275303d-e97b-bb3c-d4ae-57bd8134e6ed, 'name': SearchDatastore_Task, 'duration_secs': 0.007823} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.644894] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9e7e661-723f-4a4b-9e16-800770f0c47e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.649803] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1020.649803] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c3e331-bb3b-1567-5ad2-bbc1c923f390" [ 1020.649803] env[62552]: _type = "Task" [ 1020.649803] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.660445] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c3e331-bb3b-1567-5ad2-bbc1c923f390, 'name': SearchDatastore_Task} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.660694] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.660943] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572/462a3224-d843-45f1-a2bf-69afc18e2572.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1020.661205] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8de3de1-4cec-476e-88cd-e20422e02427 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.667598] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1020.667598] env[62552]: value = "task-1239965" [ 1020.667598] env[62552]: _type = "Task" [ 1020.667598] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.677096] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.685338] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.685600] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.685773] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.685970] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.686510] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.686510] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.686510] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.686680] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.687467] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.687467] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.687467] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.696045] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04fea705-103c-4c0a-a354-ac49c02ce86f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.710685] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1020.710685] env[62552]: value = "task-1239966" [ 1020.710685] env[62552]: _type = "Task" [ 1020.710685] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.721826] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239966, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.753194] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c1a952-a45c-41af-8372-d557238f0c0e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.760711] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508150fb-92f6-4b42-9981-90d402c5b1d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.790044] env[62552]: DEBUG nova.network.neutron [-] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.792393] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab35409-5089-4a56-a5f6-02db05479152 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.801164] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c702a722-b672-492e-accd-8d2fd7a8f036 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.817952] env[62552]: DEBUG nova.compute.provider_tree [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.859837] env[62552]: DEBUG nova.network.neutron [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Successfully created port: 84bce384-7315-4f51-9705-fba16bcbaa82 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.875456] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239964, 'name': ReconfigVM_Task, 'duration_secs': 0.227662} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.875753] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfigured VM instance instance-00000055 to detach disk 2002 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1020.876016] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1020.876468] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-731fd828-0d6f-4a26-9d30-e147d839ac47 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.884136] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1020.884136] env[62552]: value = "task-1239967" [ 1020.884136] env[62552]: _type = "Task" [ 1020.884136] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.893014] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239967, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.897357] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d4cf1d0-fe7e-4e6b-b523-9a9fbc5221ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.907482] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7e5e6d-1327-4798-8312-e0f1819c78fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.940867] env[62552]: DEBUG nova.compute.manager [req-16ab1deb-3b95-4013-bd6e-959effbe881e req-942cde9c-9398-48f2-9687-9fe18f56968d service nova] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Detach interface failed, port_id=85efda2a-fdf8-4118-861c-af85ee3f0d46, reason: Instance 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1020.960113] env[62552]: DEBUG nova.compute.manager [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1021.085597] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.086035] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.086211] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.086530] env[62552]: DEBUG oslo_concurrency.lockutils [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] Acquired lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.086720] env[62552]: DEBUG nova.network.neutron [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Refreshing network info cache for port 3688687a-8e5c-4347-8972-724ebb0df317 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1021.090433] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a305c46-4afe-4257-aff5-9bcf392fc882 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.110074] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.110810] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.110810] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.110810] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.111227] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.111227] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.111325] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.111566] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.111781] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.111995] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.112227] env[62552]: DEBUG nova.virt.hardware [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.118811] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Reconfiguring VM to attach interface {{(pid=62552) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1021.119756] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed1261ac-d90f-497d-aad6-7d13d16b7651 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.139591] env[62552]: DEBUG oslo_vmware.api [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 1021.139591] env[62552]: value = "task-1239968" [ 1021.139591] env[62552]: _type = "Task" [ 1021.139591] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.151298] env[62552]: DEBUG oslo_vmware.api [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239968, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.179086] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239965, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.222431] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239966, 'name': ReconfigVM_Task, 'duration_secs': 0.205002} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.222796] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance '88d0f502-fc3f-429e-bdf6-a1ebacec2117' progress to 33 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1021.293180] env[62552]: INFO nova.compute.manager [-] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Took 1.43 seconds to deallocate network for instance. [ 1021.322832] env[62552]: DEBUG nova.scheduler.client.report [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1021.394458] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239967, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.649518] env[62552]: DEBUG oslo_vmware.api [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239968, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.682467] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590428} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.682792] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572/462a3224-d843-45f1-a2bf-69afc18e2572.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1021.683032] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.683299] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0710a6a2-1531-47e9-b6f5-6a16c8ac7e07 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.689968] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1021.689968] env[62552]: value = "task-1239969" [ 1021.689968] env[62552]: _type = "Task" [ 1021.689968] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.700855] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239969, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.729554] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.730487] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.730487] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.730487] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.730487] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.730656] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.730900] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.731092] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.731316] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.731501] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.732085] env[62552]: DEBUG nova.virt.hardware [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.737733] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Reconfiguring VM instance instance-0000003c to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1021.738079] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcd68af2-20a5-45a1-aac7-f1039c43c169 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.759968] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1021.759968] env[62552]: value = "task-1239970" [ 1021.759968] env[62552]: _type = "Task" [ 1021.759968] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.768721] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239970, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.799874] env[62552]: DEBUG oslo_concurrency.lockutils [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.827637] env[62552]: DEBUG nova.network.neutron [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updated VIF entry in instance network info cache for port 3688687a-8e5c-4347-8972-724ebb0df317. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1021.828340] env[62552]: DEBUG nova.network.neutron [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [{"id": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "address": "fa:16:3e:c9:01:94", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34124ec0-5d", "ovs_interfaceid": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3688687a-8e5c-4347-8972-724ebb0df317", "address": "fa:16:3e:79:0b:d3", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3688687a-8e", "ovs_interfaceid": "3688687a-8e5c-4347-8972-724ebb0df317", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.830035] env[62552]: DEBUG oslo_concurrency.lockutils [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.882s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.832977] env[62552]: DEBUG oslo_concurrency.lockutils [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.033s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.833311] env[62552]: DEBUG nova.objects.instance [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'resources' on Instance uuid 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.865044] env[62552]: INFO nova.scheduler.client.report [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Deleted allocations for instance 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734 [ 1021.897459] env[62552]: DEBUG oslo_vmware.api [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1239967, 'name': PowerOnVM_Task, 'duration_secs': 0.818127} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.898734] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.898734] env[62552]: DEBUG nova.compute.manager [None req-7a1a96b8-d2c4-40b4-8564-fb009ef539e2 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1021.898992] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b26e3ca-d88a-40b1-b6ca-0403dd14cc88 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.971976] env[62552]: DEBUG nova.compute.manager [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1021.999472] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.999582] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.999735] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.999933] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1022.000097] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.000252] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1022.000463] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1022.001061] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1022.001061] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1022.001061] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1022.001178] env[62552]: DEBUG nova.virt.hardware [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1022.002124] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d14352-0773-4c83-8f86-24eb61e1ddc3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.012282] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433b76fd-56a4-4f79-a1dd-e5eb72f706f4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.150220] env[62552]: DEBUG oslo_vmware.api [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239968, 'name': ReconfigVM_Task, 'duration_secs': 0.626385} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.150765] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.151035] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Reconfigured VM to attach interface {{(pid=62552) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1022.201425] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239969, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074999} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.202458] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.202568] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8f7350-5ae2-47a7-90f1-60cc357d37ee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.225319] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572/462a3224-d843-45f1-a2bf-69afc18e2572.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.225656] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8dde96ed-bdf2-4474-b01b-667809a8506b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.247146] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1022.247146] env[62552]: value = "task-1239971" [ 1022.247146] env[62552]: _type = "Task" [ 1022.247146] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.255475] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239971, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.269210] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239970, 'name': ReconfigVM_Task, 'duration_secs': 0.332606} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.269515] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Reconfigured VM instance instance-0000003c to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1022.270370] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76990b0b-1e10-48f9-b4b6-58bf23e5e131 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.291997] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 88d0f502-fc3f-429e-bdf6-a1ebacec2117/88d0f502-fc3f-429e-bdf6-a1ebacec2117.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.292331] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c586eaf3-ba79-4e50-a073-ff3ebc3c0910 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.311092] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1022.311092] env[62552]: value = "task-1239972" [ 1022.311092] env[62552]: _type = "Task" [ 1022.311092] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.319670] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239972, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.333520] env[62552]: DEBUG oslo_concurrency.lockutils [req-892659c6-2abd-40bf-9cb9-9a3501974c71 req-0844eb2c-4ff4-4fa3-b2b8-69b22fb7f65f service nova] Releasing lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.375672] env[62552]: DEBUG oslo_concurrency.lockutils [None req-46cf90bb-a509-404b-9701-a1b4b5a81d1c tempest-ServersTestJSON-733485307 tempest-ServersTestJSON-733485307-project-member] Lock "2c8a5f30-b54f-42ff-a5e0-04eceb1ac734" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.276s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.421775] env[62552]: DEBUG nova.network.neutron [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Successfully updated port: 84bce384-7315-4f51-9705-fba16bcbaa82 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.540151] env[62552]: DEBUG nova.compute.manager [req-96a3a4a0-f624-4c1f-b8a2-ebc9ba392e1b req-cb8aaa64-75e9-4e88-aa84-ce1bde325a03 service nova] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Received event network-vif-plugged-84bce384-7315-4f51-9705-fba16bcbaa82 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1022.540151] env[62552]: DEBUG oslo_concurrency.lockutils [req-96a3a4a0-f624-4c1f-b8a2-ebc9ba392e1b req-cb8aaa64-75e9-4e88-aa84-ce1bde325a03 service nova] Acquiring lock "19b69199-b3e1-45c5-af35-d9329113d360-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.540276] env[62552]: DEBUG oslo_concurrency.lockutils [req-96a3a4a0-f624-4c1f-b8a2-ebc9ba392e1b req-cb8aaa64-75e9-4e88-aa84-ce1bde325a03 service nova] Lock "19b69199-b3e1-45c5-af35-d9329113d360-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.540435] env[62552]: DEBUG oslo_concurrency.lockutils [req-96a3a4a0-f624-4c1f-b8a2-ebc9ba392e1b req-cb8aaa64-75e9-4e88-aa84-ce1bde325a03 service nova] Lock "19b69199-b3e1-45c5-af35-d9329113d360-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.540838] env[62552]: DEBUG nova.compute.manager [req-96a3a4a0-f624-4c1f-b8a2-ebc9ba392e1b req-cb8aaa64-75e9-4e88-aa84-ce1bde325a03 service nova] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] No waiting events found dispatching network-vif-plugged-84bce384-7315-4f51-9705-fba16bcbaa82 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1022.541105] env[62552]: WARNING nova.compute.manager [req-96a3a4a0-f624-4c1f-b8a2-ebc9ba392e1b req-cb8aaa64-75e9-4e88-aa84-ce1bde325a03 service nova] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Received unexpected event network-vif-plugged-84bce384-7315-4f51-9705-fba16bcbaa82 for instance with vm_state building and task_state spawning. [ 1022.568717] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef451cb-62dd-491a-8a7b-18fc722e877f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.577024] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bf4e80-05cd-4a01-933e-1f913a44b0ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.609784] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0c4f71-842a-42e3-bd1a-04c8d2958603 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.617575] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9335c143-4559-45ad-b2e8-f202a144948d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.630543] env[62552]: DEBUG nova.compute.provider_tree [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.655988] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3198fcf9-f285-423b-9cbb-04bcf1317789 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-d0aa144e-9bcf-4faa-8d4c-7743118fbe69-3688687a-8e5c-4347-8972-724ebb0df317" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.661s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.758050] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239971, 'name': ReconfigVM_Task, 'duration_secs': 0.282314} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.758316] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572/462a3224-d843-45f1-a2bf-69afc18e2572.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.758969] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e388eab-4fb2-4e4c-88c0-7d9d3ddeb0c3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.767128] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1022.767128] env[62552]: value = "task-1239973" [ 1022.767128] env[62552]: _type = "Task" [ 1022.767128] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.776425] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239973, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.820797] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.925685] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "refresh_cache-19b69199-b3e1-45c5-af35-d9329113d360" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.925685] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired lock "refresh_cache-19b69199-b3e1-45c5-af35-d9329113d360" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.925685] env[62552]: DEBUG nova.network.neutron [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1023.040012] env[62552]: DEBUG nova.compute.manager [req-891ad76e-77d1-41cf-9c98-493d48c16acd req-77d5a9c0-d307-484d-8526-f8c060064d1c service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Received event network-changed-19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1023.040327] env[62552]: DEBUG nova.compute.manager [req-891ad76e-77d1-41cf-9c98-493d48c16acd req-77d5a9c0-d307-484d-8526-f8c060064d1c service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Refreshing instance network info cache due to event network-changed-19a87259-1f7a-428c-8790-b74e27fde6b6. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1023.040556] env[62552]: DEBUG oslo_concurrency.lockutils [req-891ad76e-77d1-41cf-9c98-493d48c16acd req-77d5a9c0-d307-484d-8526-f8c060064d1c service nova] Acquiring lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.040706] env[62552]: DEBUG oslo_concurrency.lockutils [req-891ad76e-77d1-41cf-9c98-493d48c16acd req-77d5a9c0-d307-484d-8526-f8c060064d1c service nova] Acquired lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.040869] env[62552]: DEBUG nova.network.neutron [req-891ad76e-77d1-41cf-9c98-493d48c16acd req-77d5a9c0-d307-484d-8526-f8c060064d1c service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Refreshing network info cache for port 19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1023.136019] env[62552]: DEBUG nova.scheduler.client.report [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1023.277737] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239973, 'name': Rename_Task, 'duration_secs': 0.139561} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.278190] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1023.278516] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-275e93d8-dae4-42a1-96b3-d3cf0e09de2c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.287196] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1023.287196] env[62552]: value = "task-1239974" [ 1023.287196] env[62552]: _type = "Task" [ 1023.287196] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.299823] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239974, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.323349] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.470866] env[62552]: DEBUG nova.network.neutron [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1023.638970] env[62552]: DEBUG oslo_concurrency.lockutils [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.806s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.667377] env[62552]: INFO nova.scheduler.client.report [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted allocations for instance 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5 [ 1023.724575] env[62552]: DEBUG nova.network.neutron [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Updating instance_info_cache with network_info: [{"id": "84bce384-7315-4f51-9705-fba16bcbaa82", "address": "fa:16:3e:50:fd:b6", "network": {"id": "bfa99bf2-fb08-4fb9-9231-78e1258e14a0", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-499016683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa13c44376934f519998396c182a49f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84bce384-73", "ovs_interfaceid": "84bce384-7315-4f51-9705-fba16bcbaa82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.807025] env[62552]: DEBUG oslo_vmware.api [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1239974, 'name': PowerOnVM_Task, 'duration_secs': 0.50991} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.807025] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.807025] env[62552]: INFO nova.compute.manager [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Took 8.41 seconds to spawn the instance on the hypervisor. [ 1023.807025] env[62552]: DEBUG nova.compute.manager [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1023.808828] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c625665f-b94e-41df-af07-b9c6f83e4bf6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.831475] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239972, 'name': ReconfigVM_Task, 'duration_secs': 1.323428} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.833129] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 88d0f502-fc3f-429e-bdf6-a1ebacec2117/88d0f502-fc3f-429e-bdf6-a1ebacec2117.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.833627] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance '88d0f502-fc3f-429e-bdf6-a1ebacec2117' progress to 50 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1023.977366] env[62552]: DEBUG nova.network.neutron [req-891ad76e-77d1-41cf-9c98-493d48c16acd req-77d5a9c0-d307-484d-8526-f8c060064d1c service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updated VIF entry in instance network info cache for port 19a87259-1f7a-428c-8790-b74e27fde6b6. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1023.977366] env[62552]: DEBUG nova.network.neutron [req-891ad76e-77d1-41cf-9c98-493d48c16acd req-77d5a9c0-d307-484d-8526-f8c060064d1c service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating instance_info_cache with network_info: [{"id": "19a87259-1f7a-428c-8790-b74e27fde6b6", "address": "fa:16:3e:04:c8:af", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19a87259-1f", "ovs_interfaceid": "19a87259-1f7a-428c-8790-b74e27fde6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.178329] env[62552]: DEBUG oslo_concurrency.lockutils [None req-85b83643-51f8-43c7-adbb-743776e497d4 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "2f30cf85-df40-4560-8cc3-d0cdf7cbecc5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.421s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.227514] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Releasing lock "refresh_cache-19b69199-b3e1-45c5-af35-d9329113d360" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.227861] env[62552]: DEBUG nova.compute.manager [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Instance network_info: |[{"id": "84bce384-7315-4f51-9705-fba16bcbaa82", "address": "fa:16:3e:50:fd:b6", "network": {"id": "bfa99bf2-fb08-4fb9-9231-78e1258e14a0", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-499016683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa13c44376934f519998396c182a49f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84bce384-73", "ovs_interfaceid": "84bce384-7315-4f51-9705-fba16bcbaa82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1024.228649] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:fd:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '84bce384-7315-4f51-9705-fba16bcbaa82', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.236377] env[62552]: DEBUG oslo.service.loopingcall [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.236620] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1024.236854] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-817b58c8-5b90-44c4-8090-4f9e614e696a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.262505] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.262505] env[62552]: value = "task-1239975" [ 1024.262505] env[62552]: _type = "Task" [ 1024.262505] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.269992] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239975, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.334498] env[62552]: INFO nova.compute.manager [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Took 13.58 seconds to build instance. [ 1024.344206] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "interface-d0aa144e-9bcf-4faa-8d4c-7743118fbe69-3688687a-8e5c-4347-8972-724ebb0df317" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.344206] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-d0aa144e-9bcf-4faa-8d4c-7743118fbe69-3688687a-8e5c-4347-8972-724ebb0df317" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.348196] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d172c37c-6428-412e-b825-1b04384b141c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.369546] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843291f1-f7be-4fd3-b98e-e4e6cbc6da3d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.392855] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance '88d0f502-fc3f-429e-bdf6-a1ebacec2117' progress to 67 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1024.479095] env[62552]: DEBUG oslo_concurrency.lockutils [req-891ad76e-77d1-41cf-9c98-493d48c16acd req-77d5a9c0-d307-484d-8526-f8c060064d1c service nova] Releasing lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.568997] env[62552]: DEBUG nova.compute.manager [req-f1230e06-96e8-4621-9e32-5a39bc3e5e6e req-91b9fbc4-9050-4463-8c77-8f9b3bd66a91 service nova] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Received event network-changed-84bce384-7315-4f51-9705-fba16bcbaa82 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1024.568997] env[62552]: DEBUG nova.compute.manager [req-f1230e06-96e8-4621-9e32-5a39bc3e5e6e req-91b9fbc4-9050-4463-8c77-8f9b3bd66a91 service nova] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Refreshing instance network info cache due to event network-changed-84bce384-7315-4f51-9705-fba16bcbaa82. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1024.568997] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1230e06-96e8-4621-9e32-5a39bc3e5e6e req-91b9fbc4-9050-4463-8c77-8f9b3bd66a91 service nova] Acquiring lock "refresh_cache-19b69199-b3e1-45c5-af35-d9329113d360" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.568997] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1230e06-96e8-4621-9e32-5a39bc3e5e6e req-91b9fbc4-9050-4463-8c77-8f9b3bd66a91 service nova] Acquired lock "refresh_cache-19b69199-b3e1-45c5-af35-d9329113d360" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.568997] env[62552]: DEBUG nova.network.neutron [req-f1230e06-96e8-4621-9e32-5a39bc3e5e6e req-91b9fbc4-9050-4463-8c77-8f9b3bd66a91 service nova] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Refreshing network info cache for port 84bce384-7315-4f51-9705-fba16bcbaa82 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1024.773547] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239975, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.837416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb465556-5455-4782-9e04-6161b270a09a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.092s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.852163] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.852372] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.853345] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ce902a-806d-4a23-8055-90ac80b444d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.871084] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd25e08-e829-4b18-9c16-76907944719b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.905177] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Reconfiguring VM to detach interface {{(pid=62552) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1024.909223] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6cbe023-5d46-4603-b504-2927f8e77377 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.931164] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 1024.931164] env[62552]: value = "task-1239976" [ 1024.931164] env[62552]: _type = "Task" [ 1024.931164] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.939583] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.988089] env[62552]: DEBUG nova.network.neutron [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Port 43bb3381-583a-42f4-bcd9-3db694582a1d binding to destination host cpu-1 is already ACTIVE {{(pid=62552) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1025.071126] env[62552]: DEBUG nova.compute.manager [req-6dd9e425-5d90-4825-a8e1-a8c7b09dc16e req-f03b6360-dbf9-43e2-b25b-f61a3b5bf138 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Received event network-changed-19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1025.071387] env[62552]: DEBUG nova.compute.manager [req-6dd9e425-5d90-4825-a8e1-a8c7b09dc16e req-f03b6360-dbf9-43e2-b25b-f61a3b5bf138 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Refreshing instance network info cache due to event network-changed-19a87259-1f7a-428c-8790-b74e27fde6b6. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1025.071654] env[62552]: DEBUG oslo_concurrency.lockutils [req-6dd9e425-5d90-4825-a8e1-a8c7b09dc16e req-f03b6360-dbf9-43e2-b25b-f61a3b5bf138 service nova] Acquiring lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.071835] env[62552]: DEBUG oslo_concurrency.lockutils [req-6dd9e425-5d90-4825-a8e1-a8c7b09dc16e req-f03b6360-dbf9-43e2-b25b-f61a3b5bf138 service nova] Acquired lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.072132] env[62552]: DEBUG nova.network.neutron [req-6dd9e425-5d90-4825-a8e1-a8c7b09dc16e req-f03b6360-dbf9-43e2-b25b-f61a3b5bf138 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Refreshing network info cache for port 19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1025.077404] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.077404] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.273839] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239975, 'name': CreateVM_Task, 'duration_secs': 0.594788} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.274014] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1025.274643] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.274808] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.275166] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1025.275423] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00d51e6f-3d9f-4081-9e42-e3370baffb7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.279735] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1025.279735] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52da0cc5-469d-16d3-b015-c3826fd0bfa1" [ 1025.279735] env[62552]: _type = "Task" [ 1025.279735] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.289355] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52da0cc5-469d-16d3-b015-c3826fd0bfa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.340019] env[62552]: DEBUG nova.network.neutron [req-f1230e06-96e8-4621-9e32-5a39bc3e5e6e req-91b9fbc4-9050-4463-8c77-8f9b3bd66a91 service nova] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Updated VIF entry in instance network info cache for port 84bce384-7315-4f51-9705-fba16bcbaa82. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1025.340418] env[62552]: DEBUG nova.network.neutron [req-f1230e06-96e8-4621-9e32-5a39bc3e5e6e req-91b9fbc4-9050-4463-8c77-8f9b3bd66a91 service nova] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Updating instance_info_cache with network_info: [{"id": "84bce384-7315-4f51-9705-fba16bcbaa82", "address": "fa:16:3e:50:fd:b6", "network": {"id": "bfa99bf2-fb08-4fb9-9231-78e1258e14a0", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-499016683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa13c44376934f519998396c182a49f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84bce384-73", "ovs_interfaceid": "84bce384-7315-4f51-9705-fba16bcbaa82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.441714] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.580212] env[62552]: DEBUG nova.compute.manager [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1025.786288] env[62552]: DEBUG nova.network.neutron [req-6dd9e425-5d90-4825-a8e1-a8c7b09dc16e req-f03b6360-dbf9-43e2-b25b-f61a3b5bf138 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updated VIF entry in instance network info cache for port 19a87259-1f7a-428c-8790-b74e27fde6b6. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1025.786659] env[62552]: DEBUG nova.network.neutron [req-6dd9e425-5d90-4825-a8e1-a8c7b09dc16e req-f03b6360-dbf9-43e2-b25b-f61a3b5bf138 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating instance_info_cache with network_info: [{"id": "19a87259-1f7a-428c-8790-b74e27fde6b6", "address": "fa:16:3e:04:c8:af", "network": {"id": "206b52b6-e846-4d87-8af6-6e813e450f1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-140681652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f2003df754340ab8e4cf9ca5a21b9a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19a87259-1f", "ovs_interfaceid": "19a87259-1f7a-428c-8790-b74e27fde6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.791689] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52da0cc5-469d-16d3-b015-c3826fd0bfa1, 'name': SearchDatastore_Task, 'duration_secs': 0.030456} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.792238] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.792467] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1025.792703] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.792859] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.793080] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1025.793346] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba4d3364-b523-4bcc-b376-e72f2977db6f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.801537] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1025.801720] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1025.802459] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a19b7174-a2dd-4bcc-a1c8-7f212c18c7a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.807848] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1025.807848] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e25f84-56de-4bb9-e930-3033ef5bc5b3" [ 1025.807848] env[62552]: _type = "Task" [ 1025.807848] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.815154] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e25f84-56de-4bb9-e930-3033ef5bc5b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.843250] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1230e06-96e8-4621-9e32-5a39bc3e5e6e req-91b9fbc4-9050-4463-8c77-8f9b3bd66a91 service nova] Releasing lock "refresh_cache-19b69199-b3e1-45c5-af35-d9329113d360" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.942591] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.008953] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.009222] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.009409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.103451] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.103621] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.105156] env[62552]: INFO nova.compute.claims [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.293239] env[62552]: DEBUG oslo_concurrency.lockutils [req-6dd9e425-5d90-4825-a8e1-a8c7b09dc16e req-f03b6360-dbf9-43e2-b25b-f61a3b5bf138 service nova] Releasing lock "refresh_cache-79166f2c-f864-4d8a-b1dc-e176710400cb" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.319019] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e25f84-56de-4bb9-e930-3033ef5bc5b3, 'name': SearchDatastore_Task, 'duration_secs': 0.019438} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.319179] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ed1ce08-7a0a-4ddb-8c74-6ba7cdd843c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.324493] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1026.324493] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528b2304-49cb-7ec3-3c55-ee06f4471114" [ 1026.324493] env[62552]: _type = "Task" [ 1026.324493] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.332151] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528b2304-49cb-7ec3-3c55-ee06f4471114, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.443370] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.594529] env[62552]: DEBUG nova.compute.manager [req-9ea9db69-174a-4ab2-806f-d350be19e211 req-901ea58b-d265-4330-8242-563a6db54382 service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Received event network-changed-81435491-9993-42c8-80d8-8c4f23693ce4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1026.594720] env[62552]: DEBUG nova.compute.manager [req-9ea9db69-174a-4ab2-806f-d350be19e211 req-901ea58b-d265-4330-8242-563a6db54382 service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Refreshing instance network info cache due to event network-changed-81435491-9993-42c8-80d8-8c4f23693ce4. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1026.594973] env[62552]: DEBUG oslo_concurrency.lockutils [req-9ea9db69-174a-4ab2-806f-d350be19e211 req-901ea58b-d265-4330-8242-563a6db54382 service nova] Acquiring lock "refresh_cache-462a3224-d843-45f1-a2bf-69afc18e2572" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.595217] env[62552]: DEBUG oslo_concurrency.lockutils [req-9ea9db69-174a-4ab2-806f-d350be19e211 req-901ea58b-d265-4330-8242-563a6db54382 service nova] Acquired lock "refresh_cache-462a3224-d843-45f1-a2bf-69afc18e2572" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.595430] env[62552]: DEBUG nova.network.neutron [req-9ea9db69-174a-4ab2-806f-d350be19e211 req-901ea58b-d265-4330-8242-563a6db54382 service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Refreshing network info cache for port 81435491-9993-42c8-80d8-8c4f23693ce4 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1026.718416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.718416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.718416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.718416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.718416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.720515] env[62552]: INFO nova.compute.manager [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Terminating instance [ 1026.834921] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528b2304-49cb-7ec3-3c55-ee06f4471114, 'name': SearchDatastore_Task, 'duration_secs': 0.010291} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.835155] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.835427] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 19b69199-b3e1-45c5-af35-d9329113d360/19b69199-b3e1-45c5-af35-d9329113d360.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1026.835687] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdb3caa0-8b9b-4b3c-8401-42803fee5484 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.842514] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1026.842514] env[62552]: value = "task-1239977" [ 1026.842514] env[62552]: _type = "Task" [ 1026.842514] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.851113] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239977, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.943530] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.047037] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.047280] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.047447] env[62552]: DEBUG nova.network.neutron [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.224423] env[62552]: DEBUG nova.compute.manager [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1027.224683] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1027.225595] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70062968-eda4-4e2c-8c77-e480ef6e8894 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.236341] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1027.236648] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-330cfaa0-e85b-4d67-a7d2-776630826e82 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.243472] env[62552]: DEBUG oslo_vmware.api [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1027.243472] env[62552]: value = "task-1239978" [ 1027.243472] env[62552]: _type = "Task" [ 1027.243472] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.256696] env[62552]: DEBUG oslo_vmware.api [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.313417] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733a9f7a-6d48-407a-b639-da3a12af6a65 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.322147] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab063389-1d2a-45fe-9a1d-2fe7ab697a2a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.360958] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82098af2-8f3a-4507-a41f-fe6b4e7a05df {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.371722] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239977, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.374981] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f44d70-2873-4ffa-8367-a01d15dafc38 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.393691] env[62552]: DEBUG nova.compute.provider_tree [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.444612] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.455177] env[62552]: DEBUG nova.network.neutron [req-9ea9db69-174a-4ab2-806f-d350be19e211 req-901ea58b-d265-4330-8242-563a6db54382 service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Updated VIF entry in instance network info cache for port 81435491-9993-42c8-80d8-8c4f23693ce4. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1027.455581] env[62552]: DEBUG nova.network.neutron [req-9ea9db69-174a-4ab2-806f-d350be19e211 req-901ea58b-d265-4330-8242-563a6db54382 service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Updating instance_info_cache with network_info: [{"id": "81435491-9993-42c8-80d8-8c4f23693ce4", "address": "fa:16:3e:c5:f9:cc", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81435491-99", "ovs_interfaceid": "81435491-9993-42c8-80d8-8c4f23693ce4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.754664] env[62552]: DEBUG oslo_vmware.api [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239978, 'name': PowerOffVM_Task, 'duration_secs': 0.207268} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.755531] env[62552]: DEBUG nova.network.neutron [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.756691] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1027.756872] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1027.757328] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a636da28-82b4-4b09-8d70-81d965cab68c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.814184] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1027.814523] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1027.814722] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Deleting the datastore file [datastore1] a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.814991] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8811c09a-ecdf-412f-bff1-50504991efb4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.822879] env[62552]: DEBUG oslo_vmware.api [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1027.822879] env[62552]: value = "task-1239980" [ 1027.822879] env[62552]: _type = "Task" [ 1027.822879] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.830262] env[62552]: DEBUG oslo_vmware.api [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.861769] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239977, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.750933} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.862953] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 19b69199-b3e1-45c5-af35-d9329113d360/19b69199-b3e1-45c5-af35-d9329113d360.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1027.862953] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.862953] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-81e2dead-8281-491d-996d-4ecf2d1026cd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.868694] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1027.868694] env[62552]: value = "task-1239981" [ 1027.868694] env[62552]: _type = "Task" [ 1027.868694] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.876212] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239981, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.896922] env[62552]: DEBUG nova.scheduler.client.report [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1027.944519] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.958135] env[62552]: DEBUG oslo_concurrency.lockutils [req-9ea9db69-174a-4ab2-806f-d350be19e211 req-901ea58b-d265-4330-8242-563a6db54382 service nova] Releasing lock "refresh_cache-462a3224-d843-45f1-a2bf-69afc18e2572" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.258020] env[62552]: DEBUG oslo_concurrency.lockutils [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.332612] env[62552]: DEBUG oslo_vmware.api [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.378079] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239981, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064941} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.378375] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.379175] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3f8580-706f-4eee-b456-9c7b2221fb9d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.401281] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 19b69199-b3e1-45c5-af35-d9329113d360/19b69199-b3e1-45c5-af35-d9329113d360.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.402045] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.402550] env[62552]: DEBUG nova.compute.manager [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1028.404978] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e46ba43b-eddb-4be4-ab6d-1f05bdf65455 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.426337] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1028.426337] env[62552]: value = "task-1239982" [ 1028.426337] env[62552]: _type = "Task" [ 1028.426337] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.434265] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239982, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.444724] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.779855] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed75da61-6525-4458-969f-3d99459cf92f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.798819] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7643e878-f5c9-4221-8cad-868ebe511346 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.805801] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance '88d0f502-fc3f-429e-bdf6-a1ebacec2117' progress to 83 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1028.833647] env[62552]: DEBUG oslo_vmware.api [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1239980, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.878754} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.833890] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.834086] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1028.834267] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1028.834442] env[62552]: INFO nova.compute.manager [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1028.835028] env[62552]: DEBUG oslo.service.loopingcall [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.835028] env[62552]: DEBUG nova.compute.manager [-] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1028.835028] env[62552]: DEBUG nova.network.neutron [-] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1028.921085] env[62552]: DEBUG nova.compute.utils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1028.923186] env[62552]: DEBUG nova.compute.manager [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1028.923186] env[62552]: DEBUG nova.network.neutron [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1028.938805] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.946213] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.972024] env[62552]: DEBUG nova.policy [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7f8aa8200874dddb71d8b21bd12ca04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8b8d96b464a439e9c7ef6f3e419a9bc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.261315] env[62552]: DEBUG nova.network.neutron [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Successfully created port: 69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1029.311773] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1029.312180] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6e1fa01-cee4-4349-827e-0a288735024a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.319708] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1029.319708] env[62552]: value = "task-1239983" [ 1029.319708] env[62552]: _type = "Task" [ 1029.319708] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.328469] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239983, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.427188] env[62552]: DEBUG nova.compute.manager [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1029.447907] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.455203] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239982, 'name': ReconfigVM_Task, 'duration_secs': 0.77528} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.455517] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 19b69199-b3e1-45c5-af35-d9329113d360/19b69199-b3e1-45c5-af35-d9329113d360.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.456176] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e001da3a-fc26-41fe-82b7-4d2d2a1248b4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.463128] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1029.463128] env[62552]: value = "task-1239984" [ 1029.463128] env[62552]: _type = "Task" [ 1029.463128] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.473266] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239984, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.554739] env[62552]: DEBUG nova.compute.manager [req-444d2eb8-e43f-414e-ba6d-babf85f58f89 req-a48a6d42-029d-4127-9b9d-8edcd6a04ab0 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Received event network-vif-deleted-56e2a627-1966-4c66-b733-d9e3cf3435ef {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1029.554885] env[62552]: INFO nova.compute.manager [req-444d2eb8-e43f-414e-ba6d-babf85f58f89 req-a48a6d42-029d-4127-9b9d-8edcd6a04ab0 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Neutron deleted interface 56e2a627-1966-4c66-b733-d9e3cf3435ef; detaching it from the instance and deleting it from the info cache [ 1029.554937] env[62552]: DEBUG nova.network.neutron [req-444d2eb8-e43f-414e-ba6d-babf85f58f89 req-a48a6d42-029d-4127-9b9d-8edcd6a04ab0 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.830388] env[62552]: DEBUG oslo_vmware.api [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1239983, 'name': PowerOnVM_Task, 'duration_secs': 0.406354} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.830685] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1029.830878] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-856864c0-7ea2-43fb-968a-b3cf6be44443 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance '88d0f502-fc3f-429e-bdf6-a1ebacec2117' progress to 100 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1029.948104] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.972613] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239984, 'name': Rename_Task, 'duration_secs': 0.15201} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.972884] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1029.973147] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6434e723-1bb8-468a-a57b-69509ab25a98 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.980998] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1029.980998] env[62552]: value = "task-1239985" [ 1029.980998] env[62552]: _type = "Task" [ 1029.980998] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.988220] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.031263] env[62552]: DEBUG nova.network.neutron [-] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.059032] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4df391c7-2e6d-48b9-8e82-e67758654599 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.067436] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81078f29-fdb3-46aa-bc02-702f20db023e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.099978] env[62552]: DEBUG nova.compute.manager [req-444d2eb8-e43f-414e-ba6d-babf85f58f89 req-a48a6d42-029d-4127-9b9d-8edcd6a04ab0 service nova] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Detach interface failed, port_id=56e2a627-1966-4c66-b733-d9e3cf3435ef, reason: Instance a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1030.440132] env[62552]: DEBUG nova.compute.manager [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1030.451659] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.470505] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.470764] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.470991] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.471211] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.471383] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.471536] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.471748] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.471932] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.472212] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.472400] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.472579] env[62552]: DEBUG nova.virt.hardware [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.473567] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0cb2fdd-5ef1-4c1c-918f-5af243be8b2e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.481937] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244c30cb-7e93-46a7-9491-fd4296389ada {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.501901] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239985, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.534697] env[62552]: INFO nova.compute.manager [-] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Took 1.70 seconds to deallocate network for instance. [ 1030.908288] env[62552]: DEBUG nova.compute.manager [req-9bef235e-4dad-41fd-98a1-dac5312a2766 req-53b454a1-18d1-4ca7-b63e-4ec59c946849 service nova] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Received event network-vif-plugged-69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1030.908515] env[62552]: DEBUG oslo_concurrency.lockutils [req-9bef235e-4dad-41fd-98a1-dac5312a2766 req-53b454a1-18d1-4ca7-b63e-4ec59c946849 service nova] Acquiring lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.909187] env[62552]: DEBUG oslo_concurrency.lockutils [req-9bef235e-4dad-41fd-98a1-dac5312a2766 req-53b454a1-18d1-4ca7-b63e-4ec59c946849 service nova] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.909187] env[62552]: DEBUG oslo_concurrency.lockutils [req-9bef235e-4dad-41fd-98a1-dac5312a2766 req-53b454a1-18d1-4ca7-b63e-4ec59c946849 service nova] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.909187] env[62552]: DEBUG nova.compute.manager [req-9bef235e-4dad-41fd-98a1-dac5312a2766 req-53b454a1-18d1-4ca7-b63e-4ec59c946849 service nova] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] No waiting events found dispatching network-vif-plugged-69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1030.909416] env[62552]: WARNING nova.compute.manager [req-9bef235e-4dad-41fd-98a1-dac5312a2766 req-53b454a1-18d1-4ca7-b63e-4ec59c946849 service nova] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Received unexpected event network-vif-plugged-69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f for instance with vm_state building and task_state spawning. [ 1030.954482] env[62552]: DEBUG oslo_vmware.api [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239976, 'name': ReconfigVM_Task, 'duration_secs': 5.80832} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.954739] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.954954] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Reconfigured VM to detach interface {{(pid=62552) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1030.997341] env[62552]: DEBUG oslo_vmware.api [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239985, 'name': PowerOnVM_Task, 'duration_secs': 0.781097} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.997980] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1030.997980] env[62552]: INFO nova.compute.manager [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Took 9.03 seconds to spawn the instance on the hypervisor. [ 1030.997980] env[62552]: DEBUG nova.compute.manager [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1030.998773] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c153fddc-71a1-4897-8bfe-2525cacb0b83 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.001992] env[62552]: DEBUG nova.network.neutron [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Successfully updated port: 69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.041065] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.041332] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.041555] env[62552]: DEBUG nova.objects.instance [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'resources' on Instance uuid a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.505230] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.505446] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.505446] env[62552]: DEBUG nova.network.neutron [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.516578] env[62552]: INFO nova.compute.manager [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Took 17.12 seconds to build instance. [ 1031.745080] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571bc78b-576e-4ba3-bcc2-7eba22f33a7b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.753876] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e328e189-4fe9-44a1-a91a-54f8542f2b97 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.786212] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130b157c-fdbe-4d93-b5cb-02b683e13da8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.794459] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833349c2-91c0-4023-8e81-ddf7e447dfa7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.809574] env[62552]: DEBUG nova.compute.provider_tree [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.018184] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5cbff4e4-7108-4813-b2f4-7169b89c6cfa tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "19b69199-b3e1-45c5-af35-d9329113d360" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.635s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.067609] env[62552]: DEBUG nova.network.neutron [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.295804] env[62552]: DEBUG nova.network.neutron [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance_info_cache with network_info: [{"id": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "address": "fa:16:3e:e7:15:da", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69937c5b-d6", "ovs_interfaceid": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.314626] env[62552]: DEBUG nova.scheduler.client.report [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1032.330024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.330024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquired lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.330024] env[62552]: DEBUG nova.network.neutron [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1032.512911] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.512911] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.513132] env[62552]: DEBUG nova.compute.manager [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Going to confirm migration 3 {{(pid=62552) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 1032.794277] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.794560] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.795165] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.795379] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.795560] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.797800] env[62552]: INFO nova.compute.manager [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Terminating instance [ 1032.799276] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.799540] env[62552]: DEBUG nova.compute.manager [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Instance network_info: |[{"id": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "address": "fa:16:3e:e7:15:da", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69937c5b-d6", "ovs_interfaceid": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1032.800713] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:15:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1032.808207] env[62552]: DEBUG oslo.service.loopingcall [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.809298] env[62552]: DEBUG nova.compute.manager [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1032.809500] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1032.809748] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1032.810720] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494fe782-288e-4e30-93e9-ba9e67d9aff4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.813938] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7407725-a201-424b-8665-266cdfdd8944 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.828076] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.787s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.830536] env[62552]: DEBUG nova.compute.manager [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1032.833855] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f493ff1-7048-4be3-ae2f-89a2d7db0de2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.838937] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1032.839692] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3d657d9-d6b6-4f5f-957c-975d09c15268 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.845474] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1032.845474] env[62552]: value = "task-1239986" [ 1032.845474] env[62552]: _type = "Task" [ 1032.845474] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.852299] env[62552]: DEBUG oslo_vmware.api [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 1032.852299] env[62552]: value = "task-1239987" [ 1032.852299] env[62552]: _type = "Task" [ 1032.852299] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.853299] env[62552]: INFO nova.scheduler.client.report [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Deleted allocations for instance a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0 [ 1032.863864] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239986, 'name': CreateVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.869690] env[62552]: DEBUG oslo_vmware.api [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239987, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.945201] env[62552]: DEBUG nova.compute.manager [req-f7a61c31-d6cf-411a-9e6a-1026e826f68a req-2f2f125a-8adc-4421-962e-0c207f3a8640 service nova] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Received event network-changed-69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1032.945522] env[62552]: DEBUG nova.compute.manager [req-f7a61c31-d6cf-411a-9e6a-1026e826f68a req-2f2f125a-8adc-4421-962e-0c207f3a8640 service nova] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Refreshing instance network info cache due to event network-changed-69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1032.945789] env[62552]: DEBUG oslo_concurrency.lockutils [req-f7a61c31-d6cf-411a-9e6a-1026e826f68a req-2f2f125a-8adc-4421-962e-0c207f3a8640 service nova] Acquiring lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.945955] env[62552]: DEBUG oslo_concurrency.lockutils [req-f7a61c31-d6cf-411a-9e6a-1026e826f68a req-2f2f125a-8adc-4421-962e-0c207f3a8640 service nova] Acquired lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.946143] env[62552]: DEBUG nova.network.neutron [req-f7a61c31-d6cf-411a-9e6a-1026e826f68a req-2f2f125a-8adc-4421-962e-0c207f3a8640 service nova] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Refreshing network info cache for port 69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1033.098289] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.098650] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.098862] env[62552]: DEBUG nova.network.neutron [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.099163] env[62552]: DEBUG nova.objects.instance [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'info_cache' on Instance uuid 88d0f502-fc3f-429e-bdf6-a1ebacec2117 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.125510] env[62552]: INFO nova.network.neutron [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Port 3688687a-8e5c-4347-8972-724ebb0df317 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1033.125880] env[62552]: DEBUG nova.network.neutron [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [{"id": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "address": "fa:16:3e:c9:01:94", "network": {"id": "06f1ea1a-5103-419f-a356-3b65b9a2685e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1353390070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2f57f72de0c4adc9754458f4e7040b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a3f99df-d1bc-4a37-a048-263445d4a7b0", "external-id": "nsx-vlan-transportzone-374", "segmentation_id": 374, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34124ec0-5d", "ovs_interfaceid": "34124ec0-5d30-4c38-8095-b11bda9f58fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.347434] env[62552]: INFO nova.compute.manager [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] instance snapshotting [ 1033.353986] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76d2532-30a8-4e04-aca1-f7bdc3bfbc59 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.363058] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1239986, 'name': CreateVM_Task, 'duration_secs': 0.377355} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.376822] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1033.379647] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8ff19382-09b4-421d-825f-1e17f4207569 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.662s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.385498] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.385705] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.386120] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1033.386976] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08bb250a-a32e-4aae-bbd0-8453543f2e0a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.389666] env[62552]: DEBUG oslo_vmware.api [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239987, 'name': PowerOffVM_Task, 'duration_secs': 0.212264} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.389886] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7afd986-a821-4743-9a3a-cc48f451d0f7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.391503] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1033.391681] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1033.392357] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e6024ba-add9-4d8e-b865-b20c87a53015 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.400942] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1033.400942] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527e98d3-21a7-84f0-26ba-27963d997ec5" [ 1033.400942] env[62552]: _type = "Task" [ 1033.400942] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.409414] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527e98d3-21a7-84f0-26ba-27963d997ec5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.455080] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1033.455334] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1033.455522] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleting the datastore file [datastore2] d0aa144e-9bcf-4faa-8d4c-7743118fbe69 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.455801] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f40049cc-dc8f-443a-8e52-d9af4bacd9ad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.462034] env[62552]: DEBUG oslo_vmware.api [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 1033.462034] env[62552]: value = "task-1239989" [ 1033.462034] env[62552]: _type = "Task" [ 1033.462034] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.472853] env[62552]: DEBUG oslo_vmware.api [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239989, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.628991] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Releasing lock "refresh_cache-d0aa144e-9bcf-4faa-8d4c-7743118fbe69" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.685337] env[62552]: DEBUG nova.network.neutron [req-f7a61c31-d6cf-411a-9e6a-1026e826f68a req-2f2f125a-8adc-4421-962e-0c207f3a8640 service nova] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updated VIF entry in instance network info cache for port 69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1033.685772] env[62552]: DEBUG nova.network.neutron [req-f7a61c31-d6cf-411a-9e6a-1026e826f68a req-2f2f125a-8adc-4421-962e-0c207f3a8640 service nova] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance_info_cache with network_info: [{"id": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "address": "fa:16:3e:e7:15:da", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69937c5b-d6", "ovs_interfaceid": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.902538] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1033.902935] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-32b51dfa-99d1-4fee-8c06-a934d94345d5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.914568] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527e98d3-21a7-84f0-26ba-27963d997ec5, 'name': SearchDatastore_Task, 'duration_secs': 0.012171} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.915953] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.916226] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1033.916529] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.916726] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.916947] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1033.917290] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1033.917290] env[62552]: value = "task-1239990" [ 1033.917290] env[62552]: _type = "Task" [ 1033.917290] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.917483] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74b3d05d-793f-485f-adf0-9adc0416ce35 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.927657] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239990, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.928829] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1033.929120] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1033.929951] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed8b7054-7426-4938-8a1a-7d79c7a04e4d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.935035] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1033.935035] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a1c25d-7123-1c11-f673-950cdddf71e6" [ 1033.935035] env[62552]: _type = "Task" [ 1033.935035] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.942321] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a1c25d-7123-1c11-f673-950cdddf71e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.971446] env[62552]: DEBUG oslo_vmware.api [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1239989, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217372} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.971722] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.971940] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1033.972209] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1033.972421] env[62552]: INFO nova.compute.manager [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1033.972684] env[62552]: DEBUG oslo.service.loopingcall [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.972899] env[62552]: DEBUG nova.compute.manager [-] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1033.972995] env[62552]: DEBUG nova.network.neutron [-] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1034.133483] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d060b5a9-2cd7-41ea-a01d-19140a5df677 tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "interface-d0aa144e-9bcf-4faa-8d4c-7743118fbe69-3688687a-8e5c-4347-8972-724ebb0df317" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.791s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.188892] env[62552]: DEBUG oslo_concurrency.lockutils [req-f7a61c31-d6cf-411a-9e6a-1026e826f68a req-2f2f125a-8adc-4421-962e-0c207f3a8640 service nova] Releasing lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.398707] env[62552]: DEBUG nova.network.neutron [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [{"id": "43bb3381-583a-42f4-bcd9-3db694582a1d", "address": "fa:16:3e:de:1c:09", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43bb3381-58", "ovs_interfaceid": "43bb3381-583a-42f4-bcd9-3db694582a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.428556] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239990, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.445239] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a1c25d-7123-1c11-f673-950cdddf71e6, 'name': SearchDatastore_Task, 'duration_secs': 0.008507} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.446368] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7555de9-ad83-4078-9789-37ae311eac2f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.452848] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1034.452848] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e23f37-dd55-39c1-c945-3e2484244eb6" [ 1034.452848] env[62552]: _type = "Task" [ 1034.452848] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.463624] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e23f37-dd55-39c1-c945-3e2484244eb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.836228] env[62552]: DEBUG nova.compute.manager [req-837e4ebc-325e-422c-8a5f-66e3572f293c req-2c8ab528-de8b-49cc-babe-c6d95b19945c service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Received event network-vif-deleted-34124ec0-5d30-4c38-8095-b11bda9f58fa {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1034.836228] env[62552]: INFO nova.compute.manager [req-837e4ebc-325e-422c-8a5f-66e3572f293c req-2c8ab528-de8b-49cc-babe-c6d95b19945c service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Neutron deleted interface 34124ec0-5d30-4c38-8095-b11bda9f58fa; detaching it from the instance and deleting it from the info cache [ 1034.836228] env[62552]: DEBUG nova.network.neutron [req-837e4ebc-325e-422c-8a5f-66e3572f293c req-2c8ab528-de8b-49cc-babe-c6d95b19945c service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.901965] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-88d0f502-fc3f-429e-bdf6-a1ebacec2117" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.902376] env[62552]: DEBUG nova.objects.instance [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'migration_context' on Instance uuid 88d0f502-fc3f-429e-bdf6-a1ebacec2117 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.931018] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239990, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.963340] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e23f37-dd55-39c1-c945-3e2484244eb6, 'name': SearchDatastore_Task, 'duration_secs': 0.011972} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.963838] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.963838] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] e3f9d5f3-705a-46b1-80d0-bd2421a8b074/e3f9d5f3-705a-46b1-80d0-bd2421a8b074.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1034.964125] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e27c9c76-c510-4b4b-a54b-b4e66f66faef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.971386] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1034.971386] env[62552]: value = "task-1239991" [ 1034.971386] env[62552]: _type = "Task" [ 1034.971386] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.981358] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239991, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.313636] env[62552]: DEBUG nova.network.neutron [-] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.338618] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c2dc8ce-6677-48df-b7fb-d0d83274af54 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.348843] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d063795-bfdc-4cad-a2cf-49fc33cdb82a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.380697] env[62552]: DEBUG nova.compute.manager [req-837e4ebc-325e-422c-8a5f-66e3572f293c req-2c8ab528-de8b-49cc-babe-c6d95b19945c service nova] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Detach interface failed, port_id=34124ec0-5d30-4c38-8095-b11bda9f58fa, reason: Instance d0aa144e-9bcf-4faa-8d4c-7743118fbe69 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1035.405512] env[62552]: DEBUG nova.objects.base [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Object Instance<88d0f502-fc3f-429e-bdf6-a1ebacec2117> lazy-loaded attributes: info_cache,migration_context {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1035.406509] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587db6a6-e7e7-4222-a69d-e6b59ace52b4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.429040] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d8eb747-c4da-40a7-942e-8e7fc0654763 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.434582] env[62552]: DEBUG oslo_vmware.api [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1035.434582] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cb3937-235e-419f-877f-863ad8aa7d7e" [ 1035.434582] env[62552]: _type = "Task" [ 1035.434582] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.438154] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239990, 'name': CreateSnapshot_Task, 'duration_secs': 1.162927} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.441349] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1035.442152] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1cb7b0b-d6f2-4d9e-9688-0240c4f3a43c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.455942] env[62552]: DEBUG oslo_vmware.api [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cb3937-235e-419f-877f-863ad8aa7d7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.480819] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239991, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503167} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.481109] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] e3f9d5f3-705a-46b1-80d0-bd2421a8b074/e3f9d5f3-705a-46b1-80d0-bd2421a8b074.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1035.481339] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.481599] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5975b582-271d-4bd0-96ec-c7646a4382d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.489819] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1035.489819] env[62552]: value = "task-1239992" [ 1035.489819] env[62552]: _type = "Task" [ 1035.489819] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.497255] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239992, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.816474] env[62552]: INFO nova.compute.manager [-] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Took 1.84 seconds to deallocate network for instance. [ 1035.947667] env[62552]: DEBUG oslo_vmware.api [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cb3937-235e-419f-877f-863ad8aa7d7e, 'name': SearchDatastore_Task, 'duration_secs': 0.022639} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.947998] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.948227] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.963637] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1035.964418] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-df77be66-7224-42b9-b72c-cbe57fc4ea5d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.972654] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1035.972654] env[62552]: value = "task-1239993" [ 1035.972654] env[62552]: _type = "Task" [ 1035.972654] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.980340] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239993, 'name': CloneVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.998205] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239992, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064075} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.998424] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1035.999206] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659a4f38-0cf0-41f8-82fb-72b203fc1880 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.021607] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] e3f9d5f3-705a-46b1-80d0-bd2421a8b074/e3f9d5f3-705a-46b1-80d0-bd2421a8b074.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.023209] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6306a99-a212-4be4-9cad-a1f31d2c7f6f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.042436] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1036.042436] env[62552]: value = "task-1239994" [ 1036.042436] env[62552]: _type = "Task" [ 1036.042436] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.050424] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239994, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.324712] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.332429] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.332679] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.483325] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239993, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.556433] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239994, 'name': ReconfigVM_Task, 'duration_secs': 0.316719} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.556776] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Reconfigured VM instance instance-00000064 to attach disk [datastore1] e3f9d5f3-705a-46b1-80d0-bd2421a8b074/e3f9d5f3-705a-46b1-80d0-bd2421a8b074.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.557435] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6450d686-8ec4-4cc0-a19c-56bf1fbb5da4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.563737] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1036.563737] env[62552]: value = "task-1239995" [ 1036.563737] env[62552]: _type = "Task" [ 1036.563737] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.571547] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239995, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.636244] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9950133-2363-46de-9134-c5406508f350 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.645028] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41424f7-e8cb-4889-a5c8-0f9f8c697a8a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.675418] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fccc686-5da5-41d7-81b8-ac5a2beeaf42 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.683020] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6ad64d-63fb-457f-a792-f97347156154 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.696460] env[62552]: DEBUG nova.compute.provider_tree [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.835283] env[62552]: DEBUG nova.compute.manager [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1036.884672] env[62552]: DEBUG oslo_concurrency.lockutils [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquiring lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.884966] env[62552]: DEBUG oslo_concurrency.lockutils [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.885206] env[62552]: DEBUG oslo_concurrency.lockutils [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquiring lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.885402] env[62552]: DEBUG oslo_concurrency.lockutils [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.885577] env[62552]: DEBUG oslo_concurrency.lockutils [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.887785] env[62552]: INFO nova.compute.manager [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Terminating instance [ 1036.983455] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239993, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.073299] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239995, 'name': Rename_Task, 'duration_secs': 0.362629} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.073547] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.073790] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-652c54ec-6b88-43c4-9719-83f550f103c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.079364] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1037.079364] env[62552]: value = "task-1239996" [ 1037.079364] env[62552]: _type = "Task" [ 1037.079364] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.088301] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239996, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.200472] env[62552]: DEBUG nova.scheduler.client.report [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1037.356428] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.394473] env[62552]: DEBUG nova.compute.manager [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1037.394775] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.395715] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ae6e0f-9be4-44f1-a62e-f4476e7b2456 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.403760] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.403997] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5acb13f-161f-4490-8f19-4dfc87710a24 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.411694] env[62552]: DEBUG oslo_vmware.api [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 1037.411694] env[62552]: value = "task-1239997" [ 1037.411694] env[62552]: _type = "Task" [ 1037.411694] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.420350] env[62552]: DEBUG oslo_vmware.api [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.486026] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239993, 'name': CloneVM_Task} progress is 95%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.589456] env[62552]: DEBUG oslo_vmware.api [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1239996, 'name': PowerOnVM_Task, 'duration_secs': 0.447827} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.589783] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1037.590144] env[62552]: INFO nova.compute.manager [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1037.590235] env[62552]: DEBUG nova.compute.manager [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1037.591058] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a70669c-f60b-4bc0-a177-9918e32ad0f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.921694] env[62552]: DEBUG oslo_vmware.api [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239997, 'name': PowerOffVM_Task, 'duration_secs': 0.254095} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.922226] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.922812] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.922812] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7191f1a-b6e8-4f50-9110-5b87547809d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.985506] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239993, 'name': CloneVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.987989] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.988214] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.988402] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Deleting the datastore file [datastore2] 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.988654] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db865b40-aba3-4eb9-8c13-ce7b15037419 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.993926] env[62552]: DEBUG oslo_vmware.api [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for the task: (returnval){ [ 1037.993926] env[62552]: value = "task-1239999" [ 1037.993926] env[62552]: _type = "Task" [ 1037.993926] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.000991] env[62552]: DEBUG oslo_vmware.api [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239999, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.108829] env[62552]: INFO nova.compute.manager [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Took 12.02 seconds to build instance. [ 1038.210487] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.262s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.213395] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.889s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.214052] env[62552]: DEBUG nova.objects.instance [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'resources' on Instance uuid d0aa144e-9bcf-4faa-8d4c-7743118fbe69 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.485898] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1239993, 'name': CloneVM_Task, 'duration_secs': 2.048284} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.486197] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Created linked-clone VM from snapshot [ 1038.486939] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6e4c6c-3b92-4498-bfb9-aead89812ff0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.494547] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Uploading image 41877fa0-9f76-42d2-a0d0-37c123ba350a {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1038.506014] env[62552]: DEBUG oslo_vmware.api [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239999, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.509180] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1038.509443] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f1c0e4eb-3bc0-4984-8f0b-2f49f4519faa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.515789] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1038.515789] env[62552]: value = "task-1240000" [ 1038.515789] env[62552]: _type = "Task" [ 1038.515789] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.525050] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240000, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.610893] env[62552]: DEBUG oslo_concurrency.lockutils [None req-7fa3859a-3ac8-4345-b7bb-67dc784e8db8 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.537s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.765263] env[62552]: INFO nova.scheduler.client.report [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted allocation for migration 67354786-4ad2-4dc7-8bd8-ceee01cc04a1 [ 1038.944466] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8a8696-9e4a-4ebf-9208-f3d57fde19fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.952356] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc87f107-34a0-4e33-9c0d-75fe35cad474 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.985301] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce021fba-608c-45bf-9547-119e75b49aee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.993961] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb417b6-302f-464d-9f82-90ba4a20e935 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.014805] env[62552]: DEBUG nova.compute.provider_tree [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.019581] env[62552]: DEBUG oslo_vmware.api [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Task: {'id': task-1239999, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.525104} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.023216] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.023549] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1039.023680] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1039.023868] env[62552]: INFO nova.compute.manager [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Took 1.63 seconds to destroy the instance on the hypervisor. [ 1039.024161] env[62552]: DEBUG oslo.service.loopingcall [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.024900] env[62552]: DEBUG nova.compute.manager [-] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1039.025013] env[62552]: DEBUG nova.network.neutron [-] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1039.032555] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240000, 'name': Destroy_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.270881] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50925fcb-bdd0-4ddb-9707-f8a6a0ee20f4 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.758s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.368666] env[62552]: DEBUG nova.compute.manager [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Stashing vm_state: active {{(pid=62552) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 1039.523833] env[62552]: DEBUG nova.scheduler.client.report [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1039.534387] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240000, 'name': Destroy_Task, 'duration_secs': 0.614265} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.534387] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Destroyed the VM [ 1039.534387] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1039.534387] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b7d6c7e1-bf18-417e-acdf-cdd8c7233321 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.540319] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1039.540319] env[62552]: value = "task-1240001" [ 1039.540319] env[62552]: _type = "Task" [ 1039.540319] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.549065] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240001, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.670937] env[62552]: DEBUG nova.compute.manager [req-59c0a78e-4cc5-4587-b8c8-ac3c90e94cff req-c5feeda0-0f4b-44cc-b729-4537d0ae2775 service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Received event network-vif-deleted-13ae1de6-6606-415b-9afd-0bce65b48717 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1039.671358] env[62552]: INFO nova.compute.manager [req-59c0a78e-4cc5-4587-b8c8-ac3c90e94cff req-c5feeda0-0f4b-44cc-b729-4537d0ae2775 service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Neutron deleted interface 13ae1de6-6606-415b-9afd-0bce65b48717; detaching it from the instance and deleting it from the info cache [ 1039.671525] env[62552]: DEBUG nova.network.neutron [req-59c0a78e-4cc5-4587-b8c8-ac3c90e94cff req-c5feeda0-0f4b-44cc-b729-4537d0ae2775 service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.889304] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.028500] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.815s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.031069] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.675s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.033022] env[62552]: INFO nova.compute.claims [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.051318] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240001, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.051832] env[62552]: INFO nova.scheduler.client.report [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleted allocations for instance d0aa144e-9bcf-4faa-8d4c-7743118fbe69 [ 1040.065401] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.065686] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.065905] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.066118] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.066309] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.068477] env[62552]: INFO nova.compute.manager [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Terminating instance [ 1040.148586] env[62552]: DEBUG nova.network.neutron [-] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.177041] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a54b0f4-f32d-4e58-8207-332b323bb765 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.185102] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571c7fea-d021-42bd-9b30-89f0411f7803 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.217996] env[62552]: DEBUG nova.compute.manager [req-59c0a78e-4cc5-4587-b8c8-ac3c90e94cff req-c5feeda0-0f4b-44cc-b729-4537d0ae2775 service nova] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Detach interface failed, port_id=13ae1de6-6606-415b-9afd-0bce65b48717, reason: Instance 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1040.551592] env[62552]: DEBUG oslo_vmware.api [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240001, 'name': RemoveSnapshot_Task, 'duration_secs': 0.697891} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.551827] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1040.560933] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ea3c3136-5b98-4eb6-80ee-608f636fdd1a tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "d0aa144e-9bcf-4faa-8d4c-7743118fbe69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.766s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.572042] env[62552]: DEBUG nova.compute.manager [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1040.572237] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1040.573151] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ae68d7-6ba9-493a-89cf-9cc3b7895a3f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.581400] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1040.581636] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b2ec710-f553-4f6b-a998-f58ed74b0843 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.587239] env[62552]: DEBUG oslo_vmware.api [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1040.587239] env[62552]: value = "task-1240002" [ 1040.587239] env[62552]: _type = "Task" [ 1040.587239] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.594665] env[62552]: DEBUG oslo_vmware.api [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240002, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.651520] env[62552]: INFO nova.compute.manager [-] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Took 1.63 seconds to deallocate network for instance. [ 1041.055684] env[62552]: WARNING nova.compute.manager [None req-55855f83-17c5-4a5a-aff6-ddd8204064eb tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Image not found during snapshot: nova.exception.ImageNotFound: Image 41877fa0-9f76-42d2-a0d0-37c123ba350a could not be found. [ 1041.099817] env[62552]: DEBUG oslo_vmware.api [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240002, 'name': PowerOffVM_Task, 'duration_secs': 0.330615} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.100105] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1041.100328] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1041.100600] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9127c24f-7810-4876-bed5-55fe932dcac5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.159300] env[62552]: DEBUG oslo_concurrency.lockutils [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.161364] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1041.161502] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1041.161700] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleting the datastore file [datastore1] 88d0f502-fc3f-429e-bdf6-a1ebacec2117 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.161946] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c545279-f6a6-4dde-8275-f83d24197d52 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.167806] env[62552]: DEBUG oslo_vmware.api [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1041.167806] env[62552]: value = "task-1240004" [ 1041.167806] env[62552]: _type = "Task" [ 1041.167806] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.179104] env[62552]: DEBUG oslo_vmware.api [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240004, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.217039] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0b8d37-2589-4aa6-a13b-eecd0df9cd4c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.224458] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29870641-28b1-4922-9f6c-d801b588a70c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.257543] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf488b6-ccfe-40f3-a3fb-7e8d6e0c9425 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.265023] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d86e0e6-5a30-46ea-9b82-d98a3cbec180 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.278657] env[62552]: DEBUG nova.compute.provider_tree [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.646019] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.646019] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.646019] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "6a4573f0-7152-4462-bb72-58a45b2cbd97-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.646019] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.646019] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.649873] env[62552]: INFO nova.compute.manager [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Terminating instance [ 1041.677956] env[62552]: DEBUG oslo_vmware.api [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240004, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151235} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.678251] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.678443] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1041.678628] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1041.678803] env[62552]: INFO nova.compute.manager [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1041.679055] env[62552]: DEBUG oslo.service.loopingcall [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.679254] env[62552]: DEBUG nova.compute.manager [-] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1041.679350] env[62552]: DEBUG nova.network.neutron [-] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1041.784602] env[62552]: DEBUG nova.scheduler.client.report [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1041.860188] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "19b69199-b3e1-45c5-af35-d9329113d360" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.860500] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "19b69199-b3e1-45c5-af35-d9329113d360" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.860725] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "19b69199-b3e1-45c5-af35-d9329113d360-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.860919] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "19b69199-b3e1-45c5-af35-d9329113d360-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.861104] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "19b69199-b3e1-45c5-af35-d9329113d360-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.866223] env[62552]: INFO nova.compute.manager [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Terminating instance [ 1042.153736] env[62552]: DEBUG nova.compute.manager [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1042.154077] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1042.155045] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77cf343-4cd6-48ce-b15d-d1d794a8203c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.162167] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1042.162479] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8dc1163e-2cb2-4263-8595-475d63833406 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.168466] env[62552]: DEBUG oslo_vmware.api [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 1042.168466] env[62552]: value = "task-1240005" [ 1042.168466] env[62552]: _type = "Task" [ 1042.168466] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.177518] env[62552]: DEBUG oslo_vmware.api [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1240005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.289898] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.290549] env[62552]: DEBUG nova.compute.manager [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1042.293769] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.405s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.375018] env[62552]: DEBUG nova.compute.manager [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1042.375018] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1042.375018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba541739-8177-4120-ae59-e00162192fb7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.382740] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1042.383202] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d1a2be2-8e4b-4912-a955-2c8c46500344 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.388771] env[62552]: DEBUG nova.compute.manager [req-9bf908e3-2a9a-48e1-8c7b-b8fd4044b1b4 req-e0f62455-3cc6-45fc-8eba-8794b3879387 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Received event network-vif-deleted-43bb3381-583a-42f4-bcd9-3db694582a1d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1042.389105] env[62552]: INFO nova.compute.manager [req-9bf908e3-2a9a-48e1-8c7b-b8fd4044b1b4 req-e0f62455-3cc6-45fc-8eba-8794b3879387 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Neutron deleted interface 43bb3381-583a-42f4-bcd9-3db694582a1d; detaching it from the instance and deleting it from the info cache [ 1042.389410] env[62552]: DEBUG nova.network.neutron [req-9bf908e3-2a9a-48e1-8c7b-b8fd4044b1b4 req-e0f62455-3cc6-45fc-8eba-8794b3879387 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.398068] env[62552]: DEBUG oslo_vmware.api [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1042.398068] env[62552]: value = "task-1240006" [ 1042.398068] env[62552]: _type = "Task" [ 1042.398068] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.407086] env[62552]: DEBUG oslo_vmware.api [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.679116] env[62552]: DEBUG oslo_vmware.api [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1240005, 'name': PowerOffVM_Task, 'duration_secs': 0.21401} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.679116] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1042.679116] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1042.679116] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5efb1726-2197-4b3a-a510-6e7eea19c958 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.739064] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1042.739377] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1042.739615] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleting the datastore file [datastore2] 6a4573f0-7152-4462-bb72-58a45b2cbd97 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.739897] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c441b468-c075-434f-9a6d-0cf142723b3f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.746355] env[62552]: DEBUG oslo_vmware.api [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for the task: (returnval){ [ 1042.746355] env[62552]: value = "task-1240008" [ 1042.746355] env[62552]: _type = "Task" [ 1042.746355] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.753656] env[62552]: DEBUG oslo_vmware.api [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1240008, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.799329] env[62552]: INFO nova.compute.claims [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.804282] env[62552]: DEBUG nova.compute.utils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1042.806299] env[62552]: DEBUG nova.compute.manager [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1042.806504] env[62552]: DEBUG nova.network.neutron [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1042.847915] env[62552]: DEBUG nova.policy [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21a3024f41634f18a55c4da7a2cc7548', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d97468cd7c140ed8ec157434f943d37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1042.857961] env[62552]: DEBUG nova.network.neutron [-] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.893412] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a258daeb-cce4-4c83-a018-7a92073d9632 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.905296] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03efa41-9211-46b9-821a-389b9e735e5e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.921184] env[62552]: DEBUG oslo_vmware.api [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240006, 'name': PowerOffVM_Task, 'duration_secs': 0.171734} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.921758] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1042.921995] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1042.922301] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d009564-2e23-4ee4-bf67-bb22e02e2364 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.939730] env[62552]: DEBUG nova.compute.manager [req-9bf908e3-2a9a-48e1-8c7b-b8fd4044b1b4 req-e0f62455-3cc6-45fc-8eba-8794b3879387 service nova] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Detach interface failed, port_id=43bb3381-583a-42f4-bcd9-3db694582a1d, reason: Instance 88d0f502-fc3f-429e-bdf6-a1ebacec2117 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1042.982473] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1042.982803] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1042.983123] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Deleting the datastore file [datastore2] 19b69199-b3e1-45c5-af35-d9329113d360 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.983236] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02ddd789-adbc-4df6-944f-d0ee8cf20f0b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.989411] env[62552]: DEBUG oslo_vmware.api [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1042.989411] env[62552]: value = "task-1240010" [ 1042.989411] env[62552]: _type = "Task" [ 1042.989411] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.997963] env[62552]: DEBUG oslo_vmware.api [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240010, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.050595] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "38bf968c-e757-4b46-b069-f625f5041c84" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.050830] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.202727] env[62552]: DEBUG nova.network.neutron [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Successfully created port: 28ef7f98-6d55-4c0f-9af2-e430a81bbec8 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.257085] env[62552]: DEBUG oslo_vmware.api [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Task: {'id': task-1240008, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145676} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.257085] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.257085] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1043.257240] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1043.257871] env[62552]: INFO nova.compute.manager [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1043.257871] env[62552]: DEBUG oslo.service.loopingcall [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.257871] env[62552]: DEBUG nova.compute.manager [-] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1043.257871] env[62552]: DEBUG nova.network.neutron [-] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1043.307927] env[62552]: DEBUG nova.compute.manager [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1043.312492] env[62552]: INFO nova.compute.resource_tracker [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating resource usage from migration 0cdd2c76-f597-4186-8b46-e13da9e1da53 [ 1043.360348] env[62552]: INFO nova.compute.manager [-] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Took 1.68 seconds to deallocate network for instance. [ 1043.499879] env[62552]: DEBUG oslo_vmware.api [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240010, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172148} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.500234] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.500438] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1043.500620] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1043.500828] env[62552]: INFO nova.compute.manager [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1043.501091] env[62552]: DEBUG oslo.service.loopingcall [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.501298] env[62552]: DEBUG nova.compute.manager [-] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1043.501455] env[62552]: DEBUG nova.network.neutron [-] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1043.516063] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a04699-7c41-41a1-9613-348cc4cbd467 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.525147] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe83330-cf42-4e6d-a055-59db6e628f54 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.531294] env[62552]: DEBUG nova.compute.manager [req-d1ad1205-3aba-4fbc-827e-7b1370524899 req-04923e7e-ad72-4acc-8562-4c493ed17d0e service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Received event network-vif-deleted-bb66aa2b-bd30-454a-b71b-da2a0285cef1 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1043.531496] env[62552]: INFO nova.compute.manager [req-d1ad1205-3aba-4fbc-827e-7b1370524899 req-04923e7e-ad72-4acc-8562-4c493ed17d0e service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Neutron deleted interface bb66aa2b-bd30-454a-b71b-da2a0285cef1; detaching it from the instance and deleting it from the info cache [ 1043.531671] env[62552]: DEBUG nova.network.neutron [req-d1ad1205-3aba-4fbc-827e-7b1370524899 req-04923e7e-ad72-4acc-8562-4c493ed17d0e service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.562961] env[62552]: DEBUG nova.compute.utils [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.566479] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda02530-0dc7-43e8-a8e7-1f119bfb2696 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.574884] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b05033e-9988-48bb-b3c7-5628306cd3f1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.588790] env[62552]: DEBUG nova.compute.provider_tree [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.868958] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.008050] env[62552]: DEBUG nova.network.neutron [-] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.034642] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c31fdf59-b52e-4188-bea3-4ad7cf22d9dd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.043562] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8e8aae-1363-495d-8493-222c20d43707 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.071547] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.021s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.072195] env[62552]: DEBUG nova.compute.manager [req-d1ad1205-3aba-4fbc-827e-7b1370524899 req-04923e7e-ad72-4acc-8562-4c493ed17d0e service nova] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Detach interface failed, port_id=bb66aa2b-bd30-454a-b71b-da2a0285cef1, reason: Instance 6a4573f0-7152-4462-bb72-58a45b2cbd97 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1044.091196] env[62552]: DEBUG nova.scheduler.client.report [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1044.319205] env[62552]: DEBUG nova.compute.manager [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1044.322229] env[62552]: DEBUG nova.network.neutron [-] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.349234] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.349524] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.349695] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.349884] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.350072] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.350235] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.350442] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.350607] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.350777] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.350965] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.351175] env[62552]: DEBUG nova.virt.hardware [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.352295] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6325a49-cda0-4144-9de7-9027b8113e57 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.360482] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c83f5c-c35f-4899-bc82-4bde0219f935 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.409415] env[62552]: DEBUG nova.compute.manager [req-91e91545-beaf-477b-aa8f-00cc0b4f42b4 req-7318516c-97d5-4ba8-84c0-3409636fea12 service nova] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Received event network-vif-deleted-84bce384-7315-4f51-9705-fba16bcbaa82 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1044.510747] env[62552]: INFO nova.compute.manager [-] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Took 1.25 seconds to deallocate network for instance. [ 1044.595920] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.302s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.596177] env[62552]: INFO nova.compute.manager [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Migrating [ 1044.602779] env[62552]: DEBUG oslo_concurrency.lockutils [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.444s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.603018] env[62552]: DEBUG nova.objects.instance [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lazy-loading 'resources' on Instance uuid 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.825587] env[62552]: INFO nova.compute.manager [-] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Took 1.32 seconds to deallocate network for instance. [ 1044.927361] env[62552]: DEBUG nova.network.neutron [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Successfully updated port: 28ef7f98-6d55-4c0f-9af2-e430a81bbec8 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.017702] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.112797] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.112987] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.113183] env[62552]: DEBUG nova.network.neutron [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1045.127678] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "38bf968c-e757-4b46-b069-f625f5041c84" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.127947] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.128352] env[62552]: INFO nova.compute.manager [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Attaching volume de08cb55-f95b-444c-8570-2098c4c7d744 to /dev/sdb [ 1045.162997] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c569cb3e-4177-4c86-aa9a-6845bc2ff6b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.169355] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645a20fb-b42a-4b74-8e83-aa4ea37c59c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.182717] env[62552]: DEBUG nova.virt.block_device [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Updating existing volume attachment record: 0a932793-d872-4538-8e2a-45b579b146b6 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1045.290220] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420e77fe-4acf-45d1-9f2c-2d4dda1577ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.298942] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e14742-c853-481a-8b46-8c3393bf316b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.328970] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04571783-010b-45d9-aaf7-2d803b1626fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.332363] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.337342] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6e823d-2e81-41fb-ae3e-f97868f22564 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.350689] env[62552]: DEBUG nova.compute.provider_tree [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.432039] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "refresh_cache-765bbf00-2569-4c49-9634-81d10f375175" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.432039] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired lock "refresh_cache-765bbf00-2569-4c49-9634-81d10f375175" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.432039] env[62552]: DEBUG nova.network.neutron [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1045.853723] env[62552]: DEBUG nova.scheduler.client.report [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1045.873438] env[62552]: DEBUG nova.network.neutron [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance_info_cache with network_info: [{"id": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "address": "fa:16:3e:e7:15:da", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69937c5b-d6", "ovs_interfaceid": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.971164] env[62552]: DEBUG nova.network.neutron [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1046.095816] env[62552]: DEBUG nova.network.neutron [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Updating instance_info_cache with network_info: [{"id": "28ef7f98-6d55-4c0f-9af2-e430a81bbec8", "address": "fa:16:3e:bc:51:ab", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28ef7f98-6d", "ovs_interfaceid": "28ef7f98-6d55-4c0f-9af2-e430a81bbec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.359094] env[62552]: DEBUG oslo_concurrency.lockutils [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.361438] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.493s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.361707] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.363438] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.346s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.363671] env[62552]: DEBUG nova.objects.instance [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lazy-loading 'resources' on Instance uuid 6a4573f0-7152-4462-bb72-58a45b2cbd97 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.377745] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.380989] env[62552]: INFO nova.scheduler.client.report [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Deleted allocations for instance 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9 [ 1046.385568] env[62552]: INFO nova.scheduler.client.report [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted allocations for instance 88d0f502-fc3f-429e-bdf6-a1ebacec2117 [ 1046.568670] env[62552]: DEBUG nova.compute.manager [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Received event network-vif-plugged-28ef7f98-6d55-4c0f-9af2-e430a81bbec8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1046.568670] env[62552]: DEBUG oslo_concurrency.lockutils [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.568670] env[62552]: DEBUG oslo_concurrency.lockutils [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] Lock "765bbf00-2569-4c49-9634-81d10f375175-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.568786] env[62552]: DEBUG oslo_concurrency.lockutils [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] Lock "765bbf00-2569-4c49-9634-81d10f375175-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.569886] env[62552]: DEBUG nova.compute.manager [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] No waiting events found dispatching network-vif-plugged-28ef7f98-6d55-4c0f-9af2-e430a81bbec8 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1046.569886] env[62552]: WARNING nova.compute.manager [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Received unexpected event network-vif-plugged-28ef7f98-6d55-4c0f-9af2-e430a81bbec8 for instance with vm_state building and task_state spawning. [ 1046.569886] env[62552]: DEBUG nova.compute.manager [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Received event network-changed-28ef7f98-6d55-4c0f-9af2-e430a81bbec8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1046.569886] env[62552]: DEBUG nova.compute.manager [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Refreshing instance network info cache due to event network-changed-28ef7f98-6d55-4c0f-9af2-e430a81bbec8. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1046.569886] env[62552]: DEBUG oslo_concurrency.lockutils [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] Acquiring lock "refresh_cache-765bbf00-2569-4c49-9634-81d10f375175" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.598207] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Releasing lock "refresh_cache-765bbf00-2569-4c49-9634-81d10f375175" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.598470] env[62552]: DEBUG nova.compute.manager [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Instance network_info: |[{"id": "28ef7f98-6d55-4c0f-9af2-e430a81bbec8", "address": "fa:16:3e:bc:51:ab", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28ef7f98-6d", "ovs_interfaceid": "28ef7f98-6d55-4c0f-9af2-e430a81bbec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1046.598796] env[62552]: DEBUG oslo_concurrency.lockutils [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] Acquired lock "refresh_cache-765bbf00-2569-4c49-9634-81d10f375175" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.598989] env[62552]: DEBUG nova.network.neutron [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Refreshing network info cache for port 28ef7f98-6d55-4c0f-9af2-e430a81bbec8 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1046.600315] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:51:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0636c3f6-fcb7-4954-ab07-c5cd0dee37b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '28ef7f98-6d55-4c0f-9af2-e430a81bbec8', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.609616] env[62552]: DEBUG oslo.service.loopingcall [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.611703] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1046.614671] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b51667e-fad9-495d-89f5-cefd321f3190 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.636383] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.636383] env[62552]: value = "task-1240012" [ 1046.636383] env[62552]: _type = "Task" [ 1046.636383] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.644778] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240012, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.890836] env[62552]: DEBUG oslo_concurrency.lockutils [None req-17d9100b-6fd6-4d3a-aea4-69b78efafa63 tempest-ServersTestFqdnHostnames-580530684 tempest-ServersTestFqdnHostnames-580530684-project-member] Lock "0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.006s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.895159] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0c5ff265-63ba-494a-8464-010a74bae8ba tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "88d0f502-fc3f-429e-bdf6-a1ebacec2117" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.829s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.036989] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa5a62e-e1d2-4298-bef3-2fe8e7a47838 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.044679] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8de9dd-17da-45c8-a07d-2ea339790429 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.076703] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca090295-1935-4bf9-939f-c97afaa24030 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.083824] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b5db5c-dcda-4e5e-9d2f-132ddaf36646 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.096555] env[62552]: DEBUG nova.compute.provider_tree [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.115409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "158d4227-90ca-41ae-821b-efd353928cb2" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.115647] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.115822] env[62552]: INFO nova.compute.manager [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Shelving [ 1047.147978] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240012, 'name': CreateVM_Task, 'duration_secs': 0.388332} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.148156] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1047.148859] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.149085] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.149406] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.149892] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87398969-354c-43d5-a165-0b69329abae9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.154058] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1047.154058] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528c05bd-1a21-ebfa-bda3-dcce09b20dde" [ 1047.154058] env[62552]: _type = "Task" [ 1047.154058] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.162586] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528c05bd-1a21-ebfa-bda3-dcce09b20dde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.296175] env[62552]: DEBUG nova.network.neutron [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Updated VIF entry in instance network info cache for port 28ef7f98-6d55-4c0f-9af2-e430a81bbec8. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1047.296550] env[62552]: DEBUG nova.network.neutron [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Updating instance_info_cache with network_info: [{"id": "28ef7f98-6d55-4c0f-9af2-e430a81bbec8", "address": "fa:16:3e:bc:51:ab", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28ef7f98-6d", "ovs_interfaceid": "28ef7f98-6d55-4c0f-9af2-e430a81bbec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.599370] env[62552]: DEBUG nova.scheduler.client.report [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1047.668022] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528c05bd-1a21-ebfa-bda3-dcce09b20dde, 'name': SearchDatastore_Task, 'duration_secs': 0.008869} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.668022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.668022] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.668022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.668022] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.668022] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.668022] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be762b69-61e3-47d7-a4d4-4f1bb04abb4d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.674158] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.674339] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1047.675049] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ebeafc9-e7f9-4d56-9d10-cb85eac3b35a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.680752] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1047.680752] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52814991-bd0a-6400-e2ed-3baa9fe3fbc9" [ 1047.680752] env[62552]: _type = "Task" [ 1047.680752] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.688380] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52814991-bd0a-6400-e2ed-3baa9fe3fbc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.799262] env[62552]: DEBUG oslo_concurrency.lockutils [req-b70ba06b-086c-497e-bb5f-f3c3468477dc req-9af50c10-3b75-4a96-abc5-eefb4d0be3eb service nova] Releasing lock "refresh_cache-765bbf00-2569-4c49-9634-81d10f375175" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.902018] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8229b209-955c-4f05-9f84-f6e9ade1ac31 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.923510] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance 'e3f9d5f3-705a-46b1-80d0-bd2421a8b074' progress to 0 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1048.104578] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.108616] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.774s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.108616] env[62552]: DEBUG nova.objects.instance [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lazy-loading 'resources' on Instance uuid 19b69199-b3e1-45c5-af35-d9329113d360 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.124917] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1048.125240] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-434ea83b-fb63-4589-82dd-2232a4e913ec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.133829] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1048.133829] env[62552]: value = "task-1240014" [ 1048.133829] env[62552]: _type = "Task" [ 1048.133829] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.139294] env[62552]: INFO nova.scheduler.client.report [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Deleted allocations for instance 6a4573f0-7152-4462-bb72-58a45b2cbd97 [ 1048.154937] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.195267] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52814991-bd0a-6400-e2ed-3baa9fe3fbc9, 'name': SearchDatastore_Task, 'duration_secs': 0.008229} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.196555] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d0a8fb4-e0fb-44a2-8e8a-5883dc81fe4c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.203118] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1048.203118] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fd15c6-1a75-c5dd-95ca-d4a9c2913e3e" [ 1048.203118] env[62552]: _type = "Task" [ 1048.203118] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.215035] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fd15c6-1a75-c5dd-95ca-d4a9c2913e3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.430424] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1048.430834] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f97701b-a206-445b-8aa7-ea93e045fe96 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.437688] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1048.437688] env[62552]: value = "task-1240015" [ 1048.437688] env[62552]: _type = "Task" [ 1048.437688] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.446398] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240015, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.645990] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240014, 'name': PowerOffVM_Task, 'duration_secs': 0.266392} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.648647] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.649630] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bac2e30-42be-4c7c-a023-45d7f3d047a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.672434] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215288b4-89ee-47bb-a9e3-3da6d8218d52 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.675260] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb674c82-215b-44aa-a071-f246b0ade19c tempest-AttachInterfacesTestJSON-774467094 tempest-AttachInterfacesTestJSON-774467094-project-member] Lock "6a4573f0-7152-4462-bb72-58a45b2cbd97" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.030s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.715442] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fd15c6-1a75-c5dd-95ca-d4a9c2913e3e, 'name': SearchDatastore_Task, 'duration_secs': 0.010011} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.715783] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.716154] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 765bbf00-2569-4c49-9634-81d10f375175/765bbf00-2569-4c49-9634-81d10f375175.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1048.716419] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93688dc0-fd38-427f-8162-65f80456357f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.725754] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1048.725754] env[62552]: value = "task-1240016" [ 1048.725754] env[62552]: _type = "Task" [ 1048.725754] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.734147] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240016, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.798346] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee76146-eca3-40de-ae28-62952a2292ca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.807933] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "68cb79de-e08c-4ceb-acca-e45952f67248" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.808190] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "68cb79de-e08c-4ceb-acca-e45952f67248" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.809961] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7739d0ea-68fc-4179-8a71-76c58e9787e4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.843117] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90aab1ef-03e4-485a-a5e0-085c6c21207c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.850905] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d288f0d7-75a0-47ef-9986-13e4d91fd66b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.866276] env[62552]: DEBUG nova.compute.provider_tree [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.949441] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240015, 'name': PowerOffVM_Task, 'duration_secs': 0.156711} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.949721] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.950046] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance 'e3f9d5f3-705a-46b1-80d0-bd2421a8b074' progress to 17 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1049.183327] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Creating Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1049.183723] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f07e2132-50e5-46a9-91fa-b2f1ed03a2d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.191816] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1049.191816] env[62552]: value = "task-1240017" [ 1049.191816] env[62552]: _type = "Task" [ 1049.191816] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.200869] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240017, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.235475] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240016, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458433} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.235756] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 765bbf00-2569-4c49-9634-81d10f375175/765bbf00-2569-4c49-9634-81d10f375175.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1049.235979] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.236264] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a89f7dc-f724-4e38-8cac-de1ac01bde44 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.242459] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1049.242459] env[62552]: value = "task-1240018" [ 1049.242459] env[62552]: _type = "Task" [ 1049.242459] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.250957] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240018, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.315464] env[62552]: DEBUG nova.compute.manager [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1049.369880] env[62552]: DEBUG nova.scheduler.client.report [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1049.456782] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1049.458081] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1049.458081] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1049.458081] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1049.458351] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1049.458421] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1049.458723] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1049.459538] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1049.459538] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1049.459538] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1049.459538] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1049.464884] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-253f54da-04a8-4d6a-8ab7-51e6d6f0bb3f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.481085] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1049.481085] env[62552]: value = "task-1240019" [ 1049.481085] env[62552]: _type = "Task" [ 1049.481085] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.490923] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240019, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.701779] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240017, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.730341] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1049.730610] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267512', 'volume_id': 'de08cb55-f95b-444c-8570-2098c4c7d744', 'name': 'volume-de08cb55-f95b-444c-8570-2098c4c7d744', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '38bf968c-e757-4b46-b069-f625f5041c84', 'attached_at': '', 'detached_at': '', 'volume_id': 'de08cb55-f95b-444c-8570-2098c4c7d744', 'serial': 'de08cb55-f95b-444c-8570-2098c4c7d744'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1049.731555] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b738ad6-84fc-4915-be9f-03afa786108c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.752937] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585c7cd8-1a8e-40ce-bd59-0433599a6bc3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.761923] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240018, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068427} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.775038] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.782935] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] volume-de08cb55-f95b-444c-8570-2098c4c7d744/volume-de08cb55-f95b-444c-8570-2098c4c7d744.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.786310] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5be7f0-86b2-47cb-8ed4-81d01fe37542 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.786310] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a95a0b81-3fdb-497b-861b-da641cc28316 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.822808] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 765bbf00-2569-4c49-9634-81d10f375175/765bbf00-2569-4c49-9634-81d10f375175.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.823222] env[62552]: DEBUG oslo_vmware.api [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1049.823222] env[62552]: value = "task-1240020" [ 1049.823222] env[62552]: _type = "Task" [ 1049.823222] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.825923] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-365d1c1b-1aca-475b-8b2c-e8f6890adb16 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.852397] env[62552]: DEBUG oslo_vmware.api [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240020, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.853768] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1049.853768] env[62552]: value = "task-1240021" [ 1049.853768] env[62552]: _type = "Task" [ 1049.853768] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.861631] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240021, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.862628] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.875279] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.877740] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.015s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.879572] env[62552]: INFO nova.compute.claims [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1049.900351] env[62552]: INFO nova.scheduler.client.report [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Deleted allocations for instance 19b69199-b3e1-45c5-af35-d9329113d360 [ 1049.996472] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240019, 'name': ReconfigVM_Task, 'duration_secs': 0.187186} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.996811] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance 'e3f9d5f3-705a-46b1-80d0-bd2421a8b074' progress to 33 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1050.202566] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240017, 'name': CreateSnapshot_Task, 'duration_secs': 0.752244} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.202867] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Created Snapshot of the VM instance {{(pid=62552) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1050.203676] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634f954b-80d9-451b-b4e4-d08e41b83263 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.350978] env[62552]: DEBUG oslo_vmware.api [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240020, 'name': ReconfigVM_Task, 'duration_secs': 0.527355} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.350978] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Reconfigured VM instance instance-0000005e to attach disk [datastore2] volume-de08cb55-f95b-444c-8570-2098c4c7d744/volume-de08cb55-f95b-444c-8570-2098c4c7d744.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.355229] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a54f47d-4b90-4ee0-9e09-633626682216 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.373230] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240021, 'name': ReconfigVM_Task, 'duration_secs': 0.274519} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.374563] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 765bbf00-2569-4c49-9634-81d10f375175/765bbf00-2569-4c49-9634-81d10f375175.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.375303] env[62552]: DEBUG oslo_vmware.api [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1050.375303] env[62552]: value = "task-1240022" [ 1050.375303] env[62552]: _type = "Task" [ 1050.375303] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.375511] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-def5f979-199d-474d-9791-6a1c3e49eee5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.387675] env[62552]: DEBUG oslo_vmware.api [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240022, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.389063] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1050.389063] env[62552]: value = "task-1240023" [ 1050.389063] env[62552]: _type = "Task" [ 1050.389063] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.396980] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240023, 'name': Rename_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.408911] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bf0ec8f8-783f-476c-98fd-3e57c2ef025e tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "19b69199-b3e1-45c5-af35-d9329113d360" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.548s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.503943] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1050.504209] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1050.504376] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1050.505392] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1050.505659] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1050.505865] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1050.506320] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1050.506418] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1050.506638] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1050.506842] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1050.507067] env[62552]: DEBUG nova.virt.hardware [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.512603] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1050.512985] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4fb46a9-99fd-4252-afaf-985f1240dc51 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.534559] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1050.534559] env[62552]: value = "task-1240024" [ 1050.534559] env[62552]: _type = "Task" [ 1050.534559] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.543553] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240024, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.723408] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Creating linked-clone VM from snapshot {{(pid=62552) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1050.723807] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2b164378-8319-4326-9e31-e8eb77926cb3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.732807] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1050.732807] env[62552]: value = "task-1240025" [ 1050.732807] env[62552]: _type = "Task" [ 1050.732807] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.741556] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240025, 'name': CloneVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.887805] env[62552]: DEBUG oslo_vmware.api [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240022, 'name': ReconfigVM_Task, 'duration_secs': 0.152404} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.888264] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267512', 'volume_id': 'de08cb55-f95b-444c-8570-2098c4c7d744', 'name': 'volume-de08cb55-f95b-444c-8570-2098c4c7d744', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '38bf968c-e757-4b46-b069-f625f5041c84', 'attached_at': '', 'detached_at': '', 'volume_id': 'de08cb55-f95b-444c-8570-2098c4c7d744', 'serial': 'de08cb55-f95b-444c-8570-2098c4c7d744'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1050.902474] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240023, 'name': Rename_Task, 'duration_secs': 0.157406} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.902855] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.903201] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0e2091a-88bc-439a-a390-668b4db4633e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.909285] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1050.909285] env[62552]: value = "task-1240026" [ 1050.909285] env[62552]: _type = "Task" [ 1050.909285] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.919590] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240026, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.046332] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240024, 'name': ReconfigVM_Task, 'duration_secs': 0.160999} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.046679] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1051.047610] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7750ea45-ccbb-4925-9467-781f77f29f27 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.076569] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] e3f9d5f3-705a-46b1-80d0-bd2421a8b074/e3f9d5f3-705a-46b1-80d0-bd2421a8b074.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.079580] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db4e6e71-6f93-490a-980d-31ff2966f0a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.101710] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1051.101710] env[62552]: value = "task-1240027" [ 1051.101710] env[62552]: _type = "Task" [ 1051.101710] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.107873] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e1ac97-704a-4239-923b-3c46a02a296b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.113789] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240027, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.118527] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61799329-eabf-4068-9260-e973f5447ae0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.154264] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af459323-83f3-4f85-bf4a-754787f452a0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.164134] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76897831-1d36-48d3-980c-1e9866cac2d4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.180160] env[62552]: DEBUG nova.compute.provider_tree [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.244605] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240025, 'name': CloneVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.419092] env[62552]: DEBUG oslo_vmware.api [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240026, 'name': PowerOnVM_Task, 'duration_secs': 0.452596} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.419420] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1051.419633] env[62552]: INFO nova.compute.manager [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Took 7.10 seconds to spawn the instance on the hypervisor. [ 1051.419826] env[62552]: DEBUG nova.compute.manager [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1051.420648] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39f7079-8af3-4bfe-9372-37157ab7bbf7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.525059] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.525059] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.613707] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240027, 'name': ReconfigVM_Task, 'duration_secs': 0.492173} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.613927] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Reconfigured VM instance instance-00000064 to attach disk [datastore1] e3f9d5f3-705a-46b1-80d0-bd2421a8b074/e3f9d5f3-705a-46b1-80d0-bd2421a8b074.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.614245] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance 'e3f9d5f3-705a-46b1-80d0-bd2421a8b074' progress to 50 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1051.686025] env[62552]: DEBUG nova.scheduler.client.report [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1051.743591] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240025, 'name': CloneVM_Task} progress is 95%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.939533] env[62552]: DEBUG nova.objects.instance [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'flavor' on Instance uuid 38bf968c-e757-4b46-b069-f625f5041c84 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.944871] env[62552]: INFO nova.compute.manager [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Took 14.60 seconds to build instance. [ 1052.027414] env[62552]: DEBUG nova.compute.manager [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1052.123019] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45109d5e-b556-41c3-8afd-c621cc339154 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.143032] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8f4703-0892-465a-9d80-780cf376829b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.163620] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance 'e3f9d5f3-705a-46b1-80d0-bd2421a8b074' progress to 67 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1052.189146] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.189651] env[62552]: DEBUG nova.compute.manager [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1052.245767] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240025, 'name': CloneVM_Task, 'duration_secs': 1.169029} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.246058] env[62552]: INFO nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Created linked-clone VM from snapshot [ 1052.246791] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a71a51-1249-41eb-b118-80cb298576e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.254096] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Uploading image 6b29c2df-72f4-4913-904a-2589eab9d481 {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1052.283692] env[62552]: DEBUG oslo_vmware.rw_handles [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1052.283692] env[62552]: value = "vm-267515" [ 1052.283692] env[62552]: _type = "VirtualMachine" [ 1052.283692] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1052.283981] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-55f379db-544e-440d-be5c-150596f22993 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.292416] env[62552]: DEBUG oslo_vmware.rw_handles [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lease: (returnval){ [ 1052.292416] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526c1b56-caca-62ba-1077-a2c3ae54f986" [ 1052.292416] env[62552]: _type = "HttpNfcLease" [ 1052.292416] env[62552]: } obtained for exporting VM: (result){ [ 1052.292416] env[62552]: value = "vm-267515" [ 1052.292416] env[62552]: _type = "VirtualMachine" [ 1052.292416] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1052.292931] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the lease: (returnval){ [ 1052.292931] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526c1b56-caca-62ba-1077-a2c3ae54f986" [ 1052.292931] env[62552]: _type = "HttpNfcLease" [ 1052.292931] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1052.300133] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1052.300133] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526c1b56-caca-62ba-1077-a2c3ae54f986" [ 1052.300133] env[62552]: _type = "HttpNfcLease" [ 1052.300133] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1052.315136] env[62552]: DEBUG nova.compute.manager [req-45a43bc6-dd70-442a-8387-e8b0ec72ffea req-97106d9d-d44a-4ba7-830f-1c7d09ead01e service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Received event network-changed-28ef7f98-6d55-4c0f-9af2-e430a81bbec8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1052.315136] env[62552]: DEBUG nova.compute.manager [req-45a43bc6-dd70-442a-8387-e8b0ec72ffea req-97106d9d-d44a-4ba7-830f-1c7d09ead01e service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Refreshing instance network info cache due to event network-changed-28ef7f98-6d55-4c0f-9af2-e430a81bbec8. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1052.315136] env[62552]: DEBUG oslo_concurrency.lockutils [req-45a43bc6-dd70-442a-8387-e8b0ec72ffea req-97106d9d-d44a-4ba7-830f-1c7d09ead01e service nova] Acquiring lock "refresh_cache-765bbf00-2569-4c49-9634-81d10f375175" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.315136] env[62552]: DEBUG oslo_concurrency.lockutils [req-45a43bc6-dd70-442a-8387-e8b0ec72ffea req-97106d9d-d44a-4ba7-830f-1c7d09ead01e service nova] Acquired lock "refresh_cache-765bbf00-2569-4c49-9634-81d10f375175" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.315136] env[62552]: DEBUG nova.network.neutron [req-45a43bc6-dd70-442a-8387-e8b0ec72ffea req-97106d9d-d44a-4ba7-830f-1c7d09ead01e service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Refreshing network info cache for port 28ef7f98-6d55-4c0f-9af2-e430a81bbec8 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1052.447171] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b57873c1-bd31-4584-9efc-fee711df49de tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.114s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.447756] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fbb22d7c-c73b-4f33-927a-5ddb6cbffef4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.320s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.552204] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.552640] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.554390] env[62552]: INFO nova.compute.claims [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.674146] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "38bf968c-e757-4b46-b069-f625f5041c84" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.674415] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.694908] env[62552]: DEBUG nova.compute.utils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1052.696350] env[62552]: DEBUG nova.compute.manager [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1052.696519] env[62552]: DEBUG nova.network.neutron [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1052.715114] env[62552]: DEBUG nova.network.neutron [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Port 69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f binding to destination host cpu-1 is already ACTIVE {{(pid=62552) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1052.739473] env[62552]: DEBUG nova.policy [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bdea5d6f50634b5e89852744db56db67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8495bac65967441996ecca1fd22a4da4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1052.803053] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1052.803053] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526c1b56-caca-62ba-1077-a2c3ae54f986" [ 1052.803053] env[62552]: _type = "HttpNfcLease" [ 1052.803053] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1052.803368] env[62552]: DEBUG oslo_vmware.rw_handles [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1052.803368] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526c1b56-caca-62ba-1077-a2c3ae54f986" [ 1052.803368] env[62552]: _type = "HttpNfcLease" [ 1052.803368] env[62552]: }. {{(pid=62552) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1052.804161] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f4aee5-304d-40ce-8548-f30cd3ad5b84 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.811059] env[62552]: DEBUG oslo_vmware.rw_handles [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b43dff-19cf-f16d-4916-076898642e4f/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1052.811243] env[62552]: DEBUG oslo_vmware.rw_handles [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b43dff-19cf-f16d-4916-076898642e4f/disk-0.vmdk for reading. {{(pid=62552) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1052.906538] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-19da83cc-8938-4090-a56b-999153ae3d25 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.121671] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquiring lock "6824a837-60d0-4577-927f-a54d28a68f80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.122387] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "6824a837-60d0-4577-927f-a54d28a68f80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.168159] env[62552]: DEBUG nova.network.neutron [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Successfully created port: 4f6cbf66-fc98-4868-82ef-2f234be029cb {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1053.177803] env[62552]: INFO nova.compute.manager [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Detaching volume de08cb55-f95b-444c-8570-2098c4c7d744 [ 1053.201515] env[62552]: DEBUG nova.compute.manager [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1053.229067] env[62552]: INFO nova.virt.block_device [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Attempting to driver detach volume de08cb55-f95b-444c-8570-2098c4c7d744 from mountpoint /dev/sdb [ 1053.229067] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1053.229067] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267512', 'volume_id': 'de08cb55-f95b-444c-8570-2098c4c7d744', 'name': 'volume-de08cb55-f95b-444c-8570-2098c4c7d744', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '38bf968c-e757-4b46-b069-f625f5041c84', 'attached_at': '', 'detached_at': '', 'volume_id': 'de08cb55-f95b-444c-8570-2098c4c7d744', 'serial': 'de08cb55-f95b-444c-8570-2098c4c7d744'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1053.229279] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f806c6-9174-4287-ac25-a41f17d1f58f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.255144] env[62552]: DEBUG nova.network.neutron [req-45a43bc6-dd70-442a-8387-e8b0ec72ffea req-97106d9d-d44a-4ba7-830f-1c7d09ead01e service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Updated VIF entry in instance network info cache for port 28ef7f98-6d55-4c0f-9af2-e430a81bbec8. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1053.255144] env[62552]: DEBUG nova.network.neutron [req-45a43bc6-dd70-442a-8387-e8b0ec72ffea req-97106d9d-d44a-4ba7-830f-1c7d09ead01e service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Updating instance_info_cache with network_info: [{"id": "28ef7f98-6d55-4c0f-9af2-e430a81bbec8", "address": "fa:16:3e:bc:51:ab", "network": {"id": "418d36ed-8b5e-412e-8a7d-1a20bb092ae0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1585827212-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d97468cd7c140ed8ec157434f943d37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28ef7f98-6d", "ovs_interfaceid": "28ef7f98-6d55-4c0f-9af2-e430a81bbec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.259020] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5a3214-56dc-44f7-a554-40a00a8fb606 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.265623] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c52d51-9ae1-44b4-82bf-035fec4d41fc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.288525] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1952a4ab-c2a9-4ed0-a309-4dca27ebba5c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.304990] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] The volume has not been displaced from its original location: [datastore2] volume-de08cb55-f95b-444c-8570-2098c4c7d744/volume-de08cb55-f95b-444c-8570-2098c4c7d744.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1053.310569] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Reconfiguring VM instance instance-0000005e to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1053.310998] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a912a92-b9b9-4a9b-ba50-b6a5ca3b16ca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.331366] env[62552]: DEBUG oslo_vmware.api [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1053.331366] env[62552]: value = "task-1240029" [ 1053.331366] env[62552]: _type = "Task" [ 1053.331366] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.342424] env[62552]: DEBUG oslo_vmware.api [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240029, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.627606] env[62552]: DEBUG nova.compute.manager [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1053.748714] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.748963] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.749253] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.761342] env[62552]: DEBUG oslo_concurrency.lockutils [req-45a43bc6-dd70-442a-8387-e8b0ec72ffea req-97106d9d-d44a-4ba7-830f-1c7d09ead01e service nova] Releasing lock "refresh_cache-765bbf00-2569-4c49-9634-81d10f375175" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.779727] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff64cda-acae-401c-aecc-c6a206df8cbb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.789033] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-734e1edc-7bbe-4b68-a839-250ae3e6fa79 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.828951] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7457a84-e034-4267-9b13-903e1cc7b333 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.843525] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb40351-fc5b-4ae9-8352-66542cc7bc3b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.847756] env[62552]: DEBUG oslo_vmware.api [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240029, 'name': ReconfigVM_Task, 'duration_secs': 0.227205} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.849063] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Reconfigured VM instance instance-0000005e to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1053.855324] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4346cc88-ac8d-453e-8914-be5c4fce6294 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.876218] env[62552]: DEBUG nova.compute.provider_tree [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.882522] env[62552]: DEBUG oslo_vmware.api [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1053.882522] env[62552]: value = "task-1240030" [ 1053.882522] env[62552]: _type = "Task" [ 1053.882522] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.891339] env[62552]: DEBUG oslo_vmware.api [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240030, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.151981] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.219636] env[62552]: DEBUG nova.compute.manager [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1054.244508] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.244798] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.244984] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.245201] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.245799] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.245961] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.246209] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.246457] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.246695] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.246916] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.247180] env[62552]: DEBUG nova.virt.hardware [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.248243] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e6d80b-a22d-4c90-b091-14acfd402728 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.259105] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9ffb79-12d4-4b67-889f-b9e0b83f0a63 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.379907] env[62552]: DEBUG nova.scheduler.client.report [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1054.392415] env[62552]: DEBUG oslo_vmware.api [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240030, 'name': ReconfigVM_Task, 'duration_secs': 0.145116} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.393370] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267512', 'volume_id': 'de08cb55-f95b-444c-8570-2098c4c7d744', 'name': 'volume-de08cb55-f95b-444c-8570-2098c4c7d744', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '38bf968c-e757-4b46-b069-f625f5041c84', 'attached_at': '', 'detached_at': '', 'volume_id': 'de08cb55-f95b-444c-8570-2098c4c7d744', 'serial': 'de08cb55-f95b-444c-8570-2098c4c7d744'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1054.700478] env[62552]: DEBUG nova.compute.manager [req-5d7955b1-1f11-47f4-b0f2-6ca6588c3c70 req-1faf3887-2957-43fe-993d-de7595272c7b service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Received event network-vif-plugged-4f6cbf66-fc98-4868-82ef-2f234be029cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1054.701426] env[62552]: DEBUG oslo_concurrency.lockutils [req-5d7955b1-1f11-47f4-b0f2-6ca6588c3c70 req-1faf3887-2957-43fe-993d-de7595272c7b service nova] Acquiring lock "68cb79de-e08c-4ceb-acca-e45952f67248-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.701770] env[62552]: DEBUG oslo_concurrency.lockutils [req-5d7955b1-1f11-47f4-b0f2-6ca6588c3c70 req-1faf3887-2957-43fe-993d-de7595272c7b service nova] Lock "68cb79de-e08c-4ceb-acca-e45952f67248-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.702160] env[62552]: DEBUG oslo_concurrency.lockutils [req-5d7955b1-1f11-47f4-b0f2-6ca6588c3c70 req-1faf3887-2957-43fe-993d-de7595272c7b service nova] Lock "68cb79de-e08c-4ceb-acca-e45952f67248-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.702483] env[62552]: DEBUG nova.compute.manager [req-5d7955b1-1f11-47f4-b0f2-6ca6588c3c70 req-1faf3887-2957-43fe-993d-de7595272c7b service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] No waiting events found dispatching network-vif-plugged-4f6cbf66-fc98-4868-82ef-2f234be029cb {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1054.702831] env[62552]: WARNING nova.compute.manager [req-5d7955b1-1f11-47f4-b0f2-6ca6588c3c70 req-1faf3887-2957-43fe-993d-de7595272c7b service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Received unexpected event network-vif-plugged-4f6cbf66-fc98-4868-82ef-2f234be029cb for instance with vm_state building and task_state spawning. [ 1054.797118] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.797118] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.797118] env[62552]: DEBUG nova.network.neutron [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1054.844969] env[62552]: DEBUG nova.network.neutron [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Successfully updated port: 4f6cbf66-fc98-4868-82ef-2f234be029cb {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1054.888121] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.888731] env[62552]: DEBUG nova.compute.manager [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1054.891533] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.740s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.893470] env[62552]: INFO nova.compute.claims [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.951125] env[62552]: DEBUG nova.objects.instance [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'flavor' on Instance uuid 38bf968c-e757-4b46-b069-f625f5041c84 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.347503] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.347503] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.347503] env[62552]: DEBUG nova.network.neutron [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1055.398648] env[62552]: DEBUG nova.compute.utils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1055.405260] env[62552]: DEBUG nova.compute.manager [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1055.405615] env[62552]: DEBUG nova.network.neutron [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1055.488127] env[62552]: DEBUG nova.policy [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62819c95c0f14cad89513269b1f0bf1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa13c44376934f519998396c182a49f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1055.582074] env[62552]: DEBUG nova.network.neutron [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance_info_cache with network_info: [{"id": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "address": "fa:16:3e:e7:15:da", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69937c5b-d6", "ovs_interfaceid": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.812064] env[62552]: DEBUG nova.network.neutron [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Successfully created port: f74cd14f-2def-457c-b333-023c8f6d6567 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1055.897897] env[62552]: DEBUG nova.network.neutron [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1055.903639] env[62552]: DEBUG nova.compute.manager [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1055.963486] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d6d8966e-7167-4781-bbd3-1c242bd4c465 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.288s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.088024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.119953] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca4b860-0ab1-40f7-9dee-0ed3f3d74263 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.126669] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57466268-ee80-4db0-9bed-17a5dddbee3e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.158357] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5bb60c-76a7-4b68-9d22-262c2b813b2a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.168556] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f9765c-964d-487a-9908-3bf62608e685 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.183665] env[62552]: DEBUG nova.compute.provider_tree [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.211432] env[62552]: DEBUG nova.network.neutron [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance_info_cache with network_info: [{"id": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "address": "fa:16:3e:9c:12:00", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f6cbf66-fc", "ovs_interfaceid": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.609834] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69da15fd-23ee-4f8f-8a9f-7ae7cf4d24c1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.629520] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6692f7c3-02f9-4608-8189-19ec30b59d66 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.636643] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance 'e3f9d5f3-705a-46b1-80d0-bd2421a8b074' progress to 83 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1056.687091] env[62552]: DEBUG nova.scheduler.client.report [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1056.714309] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.714647] env[62552]: DEBUG nova.compute.manager [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Instance network_info: |[{"id": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "address": "fa:16:3e:9c:12:00", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f6cbf66-fc", "ovs_interfaceid": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1056.715173] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:12:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f6cbf66-fc98-4868-82ef-2f234be029cb', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1056.722602] env[62552]: DEBUG oslo.service.loopingcall [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.723154] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1056.723510] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e99fc7a4-ec28-4767-8950-caa477f07641 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.740391] env[62552]: DEBUG nova.compute.manager [req-72164d13-9502-4fdb-aa6b-62478b3d8dc3 req-20d3253e-cc63-4287-a738-320331658373 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Received event network-changed-4f6cbf66-fc98-4868-82ef-2f234be029cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1056.740557] env[62552]: DEBUG nova.compute.manager [req-72164d13-9502-4fdb-aa6b-62478b3d8dc3 req-20d3253e-cc63-4287-a738-320331658373 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Refreshing instance network info cache due to event network-changed-4f6cbf66-fc98-4868-82ef-2f234be029cb. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1056.740818] env[62552]: DEBUG oslo_concurrency.lockutils [req-72164d13-9502-4fdb-aa6b-62478b3d8dc3 req-20d3253e-cc63-4287-a738-320331658373 service nova] Acquiring lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.740977] env[62552]: DEBUG oslo_concurrency.lockutils [req-72164d13-9502-4fdb-aa6b-62478b3d8dc3 req-20d3253e-cc63-4287-a738-320331658373 service nova] Acquired lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.741238] env[62552]: DEBUG nova.network.neutron [req-72164d13-9502-4fdb-aa6b-62478b3d8dc3 req-20d3253e-cc63-4287-a738-320331658373 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Refreshing network info cache for port 4f6cbf66-fc98-4868-82ef-2f234be029cb {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.747649] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1056.747649] env[62552]: value = "task-1240031" [ 1056.747649] env[62552]: _type = "Task" [ 1056.747649] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.756522] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240031, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.912265] env[62552]: DEBUG nova.compute.manager [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1056.941734] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1056.942070] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1056.942283] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.942597] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1056.942820] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.943033] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1056.943275] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1056.943505] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1056.943742] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1056.943972] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1056.944226] env[62552]: DEBUG nova.virt.hardware [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.945216] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c81d34-5c79-400e-aec1-6da5d289dcae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.952983] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "38bf968c-e757-4b46-b069-f625f5041c84" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.952983] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.952983] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "38bf968c-e757-4b46-b069-f625f5041c84-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.953112] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.953280] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.956155] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108a7b49-8fc6-4310-ab64-40656d1c988c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.960250] env[62552]: INFO nova.compute.manager [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Terminating instance [ 1057.144066] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1057.144335] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae282597-c3ef-46f0-bcd0-136328e2e712 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.152468] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1057.152468] env[62552]: value = "task-1240032" [ 1057.152468] env[62552]: _type = "Task" [ 1057.152468] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.161600] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240032, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.192871] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.301s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.193715] env[62552]: DEBUG nova.compute.manager [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1057.257942] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240031, 'name': CreateVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.464825] env[62552]: DEBUG nova.compute.manager [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1057.465132] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1057.466126] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e332ab-fa68-475e-9678-d216f3faf307 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.474738] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1057.475061] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78b9f16d-e7e4-4c82-9824-b444de58ed5c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.483477] env[62552]: DEBUG oslo_vmware.api [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1057.483477] env[62552]: value = "task-1240033" [ 1057.483477] env[62552]: _type = "Task" [ 1057.483477] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.492664] env[62552]: DEBUG oslo_vmware.api [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240033, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.532079] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "79166f2c-f864-4d8a-b1dc-e176710400cb" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.532397] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.652866] env[62552]: DEBUG nova.network.neutron [req-72164d13-9502-4fdb-aa6b-62478b3d8dc3 req-20d3253e-cc63-4287-a738-320331658373 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updated VIF entry in instance network info cache for port 4f6cbf66-fc98-4868-82ef-2f234be029cb. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.653256] env[62552]: DEBUG nova.network.neutron [req-72164d13-9502-4fdb-aa6b-62478b3d8dc3 req-20d3253e-cc63-4287-a738-320331658373 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance_info_cache with network_info: [{"id": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "address": "fa:16:3e:9c:12:00", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f6cbf66-fc", "ovs_interfaceid": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.665808] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240032, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.700189] env[62552]: DEBUG nova.compute.utils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1057.702444] env[62552]: DEBUG nova.compute.manager [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Not allocating networking since 'none' was specified. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1057.761426] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240031, 'name': CreateVM_Task, 'duration_secs': 0.518129} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.761614] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1057.762503] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.762715] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.763077] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1057.763376] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1792e396-ab9a-40cb-87e1-0671b8bb64d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.769320] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1057.769320] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525b8e3e-4e1b-4f55-0942-a3f4359962a1" [ 1057.769320] env[62552]: _type = "Task" [ 1057.769320] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.779129] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525b8e3e-4e1b-4f55-0942-a3f4359962a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.899263] env[62552]: DEBUG nova.network.neutron [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Successfully updated port: f74cd14f-2def-457c-b333-023c8f6d6567 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.994784] env[62552]: DEBUG oslo_vmware.api [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240033, 'name': PowerOffVM_Task, 'duration_secs': 0.309346} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.995174] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.995254] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1057.995493] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06dbe1ca-0717-4340-baaf-0fe46374b58f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.036266] env[62552]: INFO nova.compute.manager [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Detaching volume 418c407e-7b74-49e2-8014-086b46b72f2a [ 1058.068942] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1058.069612] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1058.070152] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleting the datastore file [datastore1] 38bf968c-e757-4b46-b069-f625f5041c84 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1058.070662] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21a85a9b-cbe1-40bc-9740-db4a4759a46d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.075755] env[62552]: INFO nova.virt.block_device [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Attempting to driver detach volume 418c407e-7b74-49e2-8014-086b46b72f2a from mountpoint /dev/sdb [ 1058.076008] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1058.076208] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267504', 'volume_id': '418c407e-7b74-49e2-8014-086b46b72f2a', 'name': 'volume-418c407e-7b74-49e2-8014-086b46b72f2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '79166f2c-f864-4d8a-b1dc-e176710400cb', 'attached_at': '', 'detached_at': '', 'volume_id': '418c407e-7b74-49e2-8014-086b46b72f2a', 'serial': '418c407e-7b74-49e2-8014-086b46b72f2a'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1058.077047] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75abf52f-d321-4854-93b1-2dbca024e3f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.081472] env[62552]: DEBUG oslo_vmware.api [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1058.081472] env[62552]: value = "task-1240035" [ 1058.081472] env[62552]: _type = "Task" [ 1058.081472] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.104626] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704a5969-ed2e-4aa4-a4f7-e445f8bfd940 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.111148] env[62552]: DEBUG oslo_vmware.api [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240035, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.115674] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344db4c8-8dc1-4743-8bc7-536d64ef69f0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.135942] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf1eab1-8b5c-4424-9824-99a7a0ec11b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.150857] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] The volume has not been displaced from its original location: [datastore2] volume-418c407e-7b74-49e2-8014-086b46b72f2a/volume-418c407e-7b74-49e2-8014-086b46b72f2a.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1058.156771] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfiguring VM instance instance-00000055 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1058.157132] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce7c17e0-76a5-4abb-b6b6-33ad6707851f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.172805] env[62552]: DEBUG oslo_concurrency.lockutils [req-72164d13-9502-4fdb-aa6b-62478b3d8dc3 req-20d3253e-cc63-4287-a738-320331658373 service nova] Releasing lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.178830] env[62552]: DEBUG oslo_vmware.api [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240032, 'name': PowerOnVM_Task, 'duration_secs': 0.715222} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.180165] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1058.180381] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-4aba24f6-81cf-4ecf-a4d1-b2c19880fb33 tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance 'e3f9d5f3-705a-46b1-80d0-bd2421a8b074' progress to 100 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1058.184058] env[62552]: DEBUG oslo_vmware.api [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1058.184058] env[62552]: value = "task-1240036" [ 1058.184058] env[62552]: _type = "Task" [ 1058.184058] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.192152] env[62552]: DEBUG oslo_vmware.api [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240036, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.204060] env[62552]: DEBUG nova.compute.manager [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1058.281236] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525b8e3e-4e1b-4f55-0942-a3f4359962a1, 'name': SearchDatastore_Task, 'duration_secs': 0.016865} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.281768] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.282059] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.282329] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.282552] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.282779] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.283081] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47c08693-3951-4253-9bdc-32dea241fd3f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.292042] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.292382] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1058.293294] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c18cfd82-b3a4-4d95-986e-dedc27d6ec57 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.298481] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1058.298481] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a79614-4186-c6aa-dd33-7ba6b899a32a" [ 1058.298481] env[62552]: _type = "Task" [ 1058.298481] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.306783] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a79614-4186-c6aa-dd33-7ba6b899a32a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.402532] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "refresh_cache-7e319ad7-ca42-4eb1-b303-88adbfba7bfe" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.402780] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired lock "refresh_cache-7e319ad7-ca42-4eb1-b303-88adbfba7bfe" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.403014] env[62552]: DEBUG nova.network.neutron [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1058.592524] env[62552]: DEBUG oslo_vmware.api [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240035, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.308537} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.592803] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1058.593037] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1058.593251] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1058.593432] env[62552]: INFO nova.compute.manager [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1058.593695] env[62552]: DEBUG oslo.service.loopingcall [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.593899] env[62552]: DEBUG nova.compute.manager [-] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1058.593999] env[62552]: DEBUG nova.network.neutron [-] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1058.699268] env[62552]: DEBUG oslo_vmware.api [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240036, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.765067] env[62552]: DEBUG nova.compute.manager [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Received event network-vif-plugged-f74cd14f-2def-457c-b333-023c8f6d6567 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1058.765213] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] Acquiring lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.765517] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] Lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.765689] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] Lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.765871] env[62552]: DEBUG nova.compute.manager [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] No waiting events found dispatching network-vif-plugged-f74cd14f-2def-457c-b333-023c8f6d6567 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1058.766064] env[62552]: WARNING nova.compute.manager [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Received unexpected event network-vif-plugged-f74cd14f-2def-457c-b333-023c8f6d6567 for instance with vm_state building and task_state spawning. [ 1058.766249] env[62552]: DEBUG nova.compute.manager [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Received event network-changed-f74cd14f-2def-457c-b333-023c8f6d6567 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1058.766415] env[62552]: DEBUG nova.compute.manager [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Refreshing instance network info cache due to event network-changed-f74cd14f-2def-457c-b333-023c8f6d6567. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1058.766597] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] Acquiring lock "refresh_cache-7e319ad7-ca42-4eb1-b303-88adbfba7bfe" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.809705] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a79614-4186-c6aa-dd33-7ba6b899a32a, 'name': SearchDatastore_Task, 'duration_secs': 0.010019} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.810590] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc15960a-27a8-4d6a-be06-8429cf7ab36b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.816281] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1058.816281] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526ea748-550b-7e6d-e2f5-83bef10612ed" [ 1058.816281] env[62552]: _type = "Task" [ 1058.816281] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.823754] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526ea748-550b-7e6d-e2f5-83bef10612ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.067612] env[62552]: DEBUG nova.network.neutron [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1059.205717] env[62552]: DEBUG oslo_vmware.api [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240036, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.216444] env[62552]: DEBUG nova.compute.manager [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1059.245156] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1059.245436] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1059.245598] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.245785] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1059.245937] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.249795] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1059.250120] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1059.250307] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1059.250508] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1059.250695] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1059.250876] env[62552]: DEBUG nova.virt.hardware [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1059.251840] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf5c507-05dd-4bd7-bb73-3d999427c2e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.260305] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22bb9b0-6fd4-4a12-b2eb-791e4166010a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.276451] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Instance VIF info [] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.282090] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Creating folder: Project (04641bcfcefb45979e80fc6a420f06a3). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1059.282937] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0dffa650-b942-42bd-9fcf-d3ecd0ea6bff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.294426] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Created folder: Project (04641bcfcefb45979e80fc6a420f06a3) in parent group-v267339. [ 1059.294616] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Creating folder: Instances. Parent ref: group-v267517. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1059.294871] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d0e377d-f210-4b0d-a198-0f0ec2658734 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.303369] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Created folder: Instances in parent group-v267517. [ 1059.303615] env[62552]: DEBUG oslo.service.loopingcall [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.303819] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.304069] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a9f6b63-e3e1-4c1b-9318-2cd0b7e2db1a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.320970] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.320970] env[62552]: value = "task-1240039" [ 1059.320970] env[62552]: _type = "Task" [ 1059.320970] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.328208] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526ea748-550b-7e6d-e2f5-83bef10612ed, 'name': SearchDatastore_Task, 'duration_secs': 0.029672} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.328869] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.329195] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 68cb79de-e08c-4ceb-acca-e45952f67248/68cb79de-e08c-4ceb-acca-e45952f67248.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.329486] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5526f070-e319-4940-8540-8e909d90cfe1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.334656] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240039, 'name': CreateVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.339246] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1059.339246] env[62552]: value = "task-1240040" [ 1059.339246] env[62552]: _type = "Task" [ 1059.339246] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.355910] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240040, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.700541] env[62552]: DEBUG oslo_vmware.api [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240036, 'name': ReconfigVM_Task, 'duration_secs': 1.367779} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.701902] env[62552]: DEBUG nova.network.neutron [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Updating instance_info_cache with network_info: [{"id": "f74cd14f-2def-457c-b333-023c8f6d6567", "address": "fa:16:3e:75:38:fa", "network": {"id": "bfa99bf2-fb08-4fb9-9231-78e1258e14a0", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-499016683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa13c44376934f519998396c182a49f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf74cd14f-2d", "ovs_interfaceid": "f74cd14f-2def-457c-b333-023c8f6d6567", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.704730] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Reconfigured VM instance instance-00000055 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1059.719254] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-092e4b7e-e06c-493a-ae2b-e85ea18ec2f0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.745487] env[62552]: DEBUG oslo_vmware.api [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1059.745487] env[62552]: value = "task-1240041" [ 1059.745487] env[62552]: _type = "Task" [ 1059.745487] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.759669] env[62552]: DEBUG oslo_vmware.api [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240041, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.835098] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240039, 'name': CreateVM_Task, 'duration_secs': 0.315127} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.835694] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1059.836644] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.836834] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.837251] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.837743] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-876d7bd3-cc7c-4966-a490-75f9c1787de3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.850918] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240040, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.851280] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1059.851280] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521b696d-c943-9881-6b84-f773d6cee498" [ 1059.851280] env[62552]: _type = "Task" [ 1059.851280] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.860521] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521b696d-c943-9881-6b84-f773d6cee498, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.069033] env[62552]: DEBUG nova.network.neutron [-] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.211685] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Releasing lock "refresh_cache-7e319ad7-ca42-4eb1-b303-88adbfba7bfe" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.212021] env[62552]: DEBUG nova.compute.manager [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Instance network_info: |[{"id": "f74cd14f-2def-457c-b333-023c8f6d6567", "address": "fa:16:3e:75:38:fa", "network": {"id": "bfa99bf2-fb08-4fb9-9231-78e1258e14a0", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-499016683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa13c44376934f519998396c182a49f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf74cd14f-2d", "ovs_interfaceid": "f74cd14f-2def-457c-b333-023c8f6d6567", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1060.212354] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] Acquired lock "refresh_cache-7e319ad7-ca42-4eb1-b303-88adbfba7bfe" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.212600] env[62552]: DEBUG nova.network.neutron [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Refreshing network info cache for port f74cd14f-2def-457c-b333-023c8f6d6567 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1060.213817] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:38:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f74cd14f-2def-457c-b333-023c8f6d6567', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1060.221991] env[62552]: DEBUG oslo.service.loopingcall [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.225016] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1060.225540] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b5186e6-26cd-441f-a08f-8bdd27b3f43d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.246062] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1060.246062] env[62552]: value = "task-1240042" [ 1060.246062] env[62552]: _type = "Task" [ 1060.246062] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.257575] env[62552]: DEBUG oslo_vmware.api [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240041, 'name': ReconfigVM_Task, 'duration_secs': 0.272503} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.260916] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267504', 'volume_id': '418c407e-7b74-49e2-8014-086b46b72f2a', 'name': 'volume-418c407e-7b74-49e2-8014-086b46b72f2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '79166f2c-f864-4d8a-b1dc-e176710400cb', 'attached_at': '', 'detached_at': '', 'volume_id': '418c407e-7b74-49e2-8014-086b46b72f2a', 'serial': '418c407e-7b74-49e2-8014-086b46b72f2a'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1060.263020] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240042, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.320561] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.320826] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.321020] env[62552]: DEBUG nova.compute.manager [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Going to confirm migration 4 {{(pid=62552) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 1060.350708] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240040, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597871} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.351036] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 68cb79de-e08c-4ceb-acca-e45952f67248/68cb79de-e08c-4ceb-acca-e45952f67248.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1060.351291] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.351600] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-183cbb10-475b-4730-b477-ea22ce6691be {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.365016] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521b696d-c943-9881-6b84-f773d6cee498, 'name': SearchDatastore_Task, 'duration_secs': 0.053538} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.366263] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.366908] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.366908] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.367099] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.367316] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.367645] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1060.367645] env[62552]: value = "task-1240043" [ 1060.367645] env[62552]: _type = "Task" [ 1060.367645] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.368153] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-376c064b-e30f-4712-a83d-fd962322687c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.381745] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.383151] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.383339] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1060.384153] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-404dea50-8b4a-401e-b595-4ae027aaf758 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.389717] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1060.389717] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5247d0ad-ebfa-cb8a-e25c-3152b8f7ad63" [ 1060.389717] env[62552]: _type = "Task" [ 1060.389717] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.401735] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5247d0ad-ebfa-cb8a-e25c-3152b8f7ad63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.483546] env[62552]: DEBUG nova.network.neutron [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Updated VIF entry in instance network info cache for port f74cd14f-2def-457c-b333-023c8f6d6567. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1060.484170] env[62552]: DEBUG nova.network.neutron [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Updating instance_info_cache with network_info: [{"id": "f74cd14f-2def-457c-b333-023c8f6d6567", "address": "fa:16:3e:75:38:fa", "network": {"id": "bfa99bf2-fb08-4fb9-9231-78e1258e14a0", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-499016683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa13c44376934f519998396c182a49f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf74cd14f-2d", "ovs_interfaceid": "f74cd14f-2def-457c-b333-023c8f6d6567", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.572242] env[62552]: INFO nova.compute.manager [-] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Took 1.98 seconds to deallocate network for instance. [ 1060.759597] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240042, 'name': CreateVM_Task, 'duration_secs': 0.365185} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.759777] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.760530] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.760704] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.761044] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.761309] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27aa73ea-fa9e-4e8a-b999-a4e40d625764 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.766114] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1060.766114] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525933c1-9dd3-021b-7576-d61d98fc04a2" [ 1060.766114] env[62552]: _type = "Task" [ 1060.766114] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.773810] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525933c1-9dd3-021b-7576-d61d98fc04a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.804953] env[62552]: DEBUG nova.objects.instance [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lazy-loading 'flavor' on Instance uuid 79166f2c-f864-4d8a-b1dc-e176710400cb {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.852210] env[62552]: DEBUG nova.compute.manager [req-f3155d8b-6203-4240-bb6b-7732c13662c9 req-737c94ea-c4a2-45a3-801e-218678c1ab9c service nova] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Received event network-vif-deleted-9cf23ecf-7022-40a4-9182-2ac0d24f6fe4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1060.857102] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.857445] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquired lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.857445] env[62552]: DEBUG nova.network.neutron [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1060.857540] env[62552]: DEBUG nova.objects.instance [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'info_cache' on Instance uuid e3f9d5f3-705a-46b1-80d0-bd2421a8b074 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.880187] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072102} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.880467] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1060.881246] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c526a48-f662-49bc-9996-97ca5c737c5b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.903883] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 68cb79de-e08c-4ceb-acca-e45952f67248/68cb79de-e08c-4ceb-acca-e45952f67248.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.907144] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fcd659c-a047-4654-93db-4436b54dc335 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.928293] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5247d0ad-ebfa-cb8a-e25c-3152b8f7ad63, 'name': SearchDatastore_Task, 'duration_secs': 0.010506} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.930420] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1060.930420] env[62552]: value = "task-1240044" [ 1060.930420] env[62552]: _type = "Task" [ 1060.930420] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.930649] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf950445-2517-4782-a068-ec6799e92f22 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.939628] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1060.939628] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526212cd-4f46-e916-01fc-ca7f71f47181" [ 1060.939628] env[62552]: _type = "Task" [ 1060.939628] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.942594] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240044, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.949664] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526212cd-4f46-e916-01fc-ca7f71f47181, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.987438] env[62552]: DEBUG oslo_concurrency.lockutils [req-f1fb66eb-668e-4174-86d2-fa8e5082741e req-cb9f2f13-a099-4881-b416-a75eef50438a service nova] Releasing lock "refresh_cache-7e319ad7-ca42-4eb1-b303-88adbfba7bfe" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.079470] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.079826] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.080020] env[62552]: DEBUG nova.objects.instance [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'resources' on Instance uuid 38bf968c-e757-4b46-b069-f625f5041c84 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.278288] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525933c1-9dd3-021b-7576-d61d98fc04a2, 'name': SearchDatastore_Task, 'duration_secs': 0.05006} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.278609] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.278868] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.279135] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.443752] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240044, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.451835] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]526212cd-4f46-e916-01fc-ca7f71f47181, 'name': SearchDatastore_Task, 'duration_secs': 0.01997} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.452108] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.452373] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6824a837-60d0-4577-927f-a54d28a68f80/6824a837-60d0-4577-927f-a54d28a68f80.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1061.452684] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.452878] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.453110] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbecabaa-c366-4b07-8208-c656957830f3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.455249] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-655d2020-34c5-4fa3-9507-8cee82449d8a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.461649] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1061.461649] env[62552]: value = "task-1240045" [ 1061.461649] env[62552]: _type = "Task" [ 1061.461649] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.465670] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.465860] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1061.467012] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bffa186-31aa-438b-83b1-9779ee05c1e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.472244] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240045, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.475156] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1061.475156] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524ff350-e69a-f281-d56f-b5cf7f0291ae" [ 1061.475156] env[62552]: _type = "Task" [ 1061.475156] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.484644] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524ff350-e69a-f281-d56f-b5cf7f0291ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.774270] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52492171-44bf-4039-b672-f628611683c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.782816] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1031c787-32cc-4ef2-b184-85b59306d197 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.813565] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6a22b408-5fcb-4aff-8629-159d7c6fecb4 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.281s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.815522] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d04736b-bfca-443a-a6a0-3086a3e66c4b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.824711] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b3b940-ee7f-491c-b0e7-576a8f347ee1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.841175] env[62552]: DEBUG nova.compute.provider_tree [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.857051] env[62552]: DEBUG oslo_concurrency.lockutils [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "462a3224-d843-45f1-a2bf-69afc18e2572" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.857401] env[62552]: DEBUG oslo_concurrency.lockutils [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.943277] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240044, 'name': ReconfigVM_Task, 'duration_secs': 0.826558} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.945912] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 68cb79de-e08c-4ceb-acca-e45952f67248/68cb79de-e08c-4ceb-acca-e45952f67248.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.946611] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fabf353-7e85-4a5c-8f85-f7ae18f61a3e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.953466] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1061.953466] env[62552]: value = "task-1240046" [ 1061.953466] env[62552]: _type = "Task" [ 1061.953466] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.962922] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240046, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.973918] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240045, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.985753] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]524ff350-e69a-f281-d56f-b5cf7f0291ae, 'name': SearchDatastore_Task, 'duration_secs': 0.016309} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.986600] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcc37abd-5ec6-447f-bd42-234c548aa0ee {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.994557] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1061.994557] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b4b31a-a945-8f98-7272-1e2fa9572949" [ 1061.994557] env[62552]: _type = "Task" [ 1061.994557] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.002336] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b4b31a-a945-8f98-7272-1e2fa9572949, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.088846] env[62552]: DEBUG nova.network.neutron [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance_info_cache with network_info: [{"id": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "address": "fa:16:3e:e7:15:da", "network": {"id": "02307bef-786b-4576-b91c-af74d1c2d3a0", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1321129108-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8b8d96b464a439e9c7ef6f3e419a9bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9e6b7d9f-c4e9-4623-9eb5-840ca1a8224c", "external-id": "nsx-vlan-transportzone-782", "segmentation_id": 782, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69937c5b-d6", "ovs_interfaceid": "69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.176889] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.177158] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.177288] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1062.345177] env[62552]: DEBUG nova.scheduler.client.report [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1062.360377] env[62552]: DEBUG nova.compute.utils [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.463275] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240046, 'name': Rename_Task, 'duration_secs': 0.365901} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.466451] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1062.466723] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-880fe8e7-2fbf-49e6-8d63-5822990a09d4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.473670] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240045, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546281} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.474825] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6824a837-60d0-4577-927f-a54d28a68f80/6824a837-60d0-4577-927f-a54d28a68f80.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1062.475062] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1062.475364] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1062.475364] env[62552]: value = "task-1240047" [ 1062.475364] env[62552]: _type = "Task" [ 1062.475364] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.475553] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff41721a-45a8-45a3-9fa8-41e8d231fd0f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.484952] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240047, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.486130] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1062.486130] env[62552]: value = "task-1240048" [ 1062.486130] env[62552]: _type = "Task" [ 1062.486130] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.497101] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240048, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.508728] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b4b31a-a945-8f98-7272-1e2fa9572949, 'name': SearchDatastore_Task, 'duration_secs': 0.009744} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.508989] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.509279] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 7e319ad7-ca42-4eb1-b303-88adbfba7bfe/7e319ad7-ca42-4eb1-b303-88adbfba7bfe.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1062.509535] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39d6e0ac-3f15-417e-8010-d2307268c695 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.515395] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1062.515395] env[62552]: value = "task-1240049" [ 1062.515395] env[62552]: _type = "Task" [ 1062.515395] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.524019] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240049, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.592563] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Releasing lock "refresh_cache-e3f9d5f3-705a-46b1-80d0-bd2421a8b074" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.593009] env[62552]: DEBUG nova.objects.instance [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'migration_context' on Instance uuid e3f9d5f3-705a-46b1-80d0-bd2421a8b074 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.607540] env[62552]: DEBUG oslo_vmware.rw_handles [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b43dff-19cf-f16d-4916-076898642e4f/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1062.608474] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8764f012-58c9-45ae-99fc-1527c7840764 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.615112] env[62552]: DEBUG oslo_vmware.rw_handles [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b43dff-19cf-f16d-4916-076898642e4f/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1062.615315] env[62552]: ERROR oslo_vmware.rw_handles [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b43dff-19cf-f16d-4916-076898642e4f/disk-0.vmdk due to incomplete transfer. [ 1062.615543] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5cce55d2-91a4-4010-867c-0b579e74113d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.622017] env[62552]: DEBUG oslo_vmware.rw_handles [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b43dff-19cf-f16d-4916-076898642e4f/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1062.622238] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Uploaded image 6b29c2df-72f4-4913-904a-2589eab9d481 to the Glance image server {{(pid=62552) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1062.625071] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Destroying the VM {{(pid=62552) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1062.625308] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f5e934c0-4fc4-4c4f-be99-17bfa367a1f5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.630276] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1062.630276] env[62552]: value = "task-1240050" [ 1062.630276] env[62552]: _type = "Task" [ 1062.630276] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.637583] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240050, 'name': Destroy_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.850396] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.863024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.874140] env[62552]: INFO nova.scheduler.client.report [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleted allocations for instance 38bf968c-e757-4b46-b069-f625f5041c84 [ 1062.986493] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240047, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.993887] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240048, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070362} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.994162] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.994917] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b63d274-f1cd-434f-abc3-2221eae3bd0e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.014297] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 6824a837-60d0-4577-927f-a54d28a68f80/6824a837-60d0-4577-927f-a54d28a68f80.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.014567] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d22605d8-adb0-431a-8787-2a58e09584e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.036560] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240049, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445894} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.037672] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 7e319ad7-ca42-4eb1-b303-88adbfba7bfe/7e319ad7-ca42-4eb1-b303-88adbfba7bfe.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1063.037893] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.038203] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1063.038203] env[62552]: value = "task-1240051" [ 1063.038203] env[62552]: _type = "Task" [ 1063.038203] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.038393] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-73e941d6-d422-4f23-8381-b085c2adedf8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.047751] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240051, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.048913] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1063.048913] env[62552]: value = "task-1240052" [ 1063.048913] env[62552]: _type = "Task" [ 1063.048913] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.056264] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240052, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.095908] env[62552]: DEBUG nova.objects.base [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1063.096917] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745d09e5-4e6a-426f-bb3e-1f0f9b829dae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.117600] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aad0e664-424d-4048-9aae-df0ad997a0ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.123278] env[62552]: DEBUG oslo_vmware.api [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1063.123278] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b81871-1511-8707-44bd-3270afe95d1a" [ 1063.123278] env[62552]: _type = "Task" [ 1063.123278] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.131201] env[62552]: DEBUG oslo_vmware.api [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b81871-1511-8707-44bd-3270afe95d1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.138257] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240050, 'name': Destroy_Task} progress is 33%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.383150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a6cdcbfd-5a55-4825-8408-d375567de98c tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "38bf968c-e757-4b46-b069-f625f5041c84" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.430s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.488015] env[62552]: DEBUG oslo_vmware.api [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240047, 'name': PowerOnVM_Task, 'duration_secs': 0.851548} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.488326] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1063.488563] env[62552]: INFO nova.compute.manager [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Took 9.27 seconds to spawn the instance on the hypervisor. [ 1063.488774] env[62552]: DEBUG nova.compute.manager [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1063.489633] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5422fd05-ae1f-44d0-bbf4-78354a0f0d28 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.551769] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240051, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.560317] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240052, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064686} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.560623] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.562059] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f2aeed-6daa-4bda-91a2-8b1ac450e1b2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.585081] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 7e319ad7-ca42-4eb1-b303-88adbfba7bfe/7e319ad7-ca42-4eb1-b303-88adbfba7bfe.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.585381] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-822fe000-3f52-41a2-aff4-5695c464ad8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.604692] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1063.604692] env[62552]: value = "task-1240053" [ 1063.604692] env[62552]: _type = "Task" [ 1063.604692] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.612596] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240053, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.632571] env[62552]: DEBUG oslo_vmware.api [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b81871-1511-8707-44bd-3270afe95d1a, 'name': SearchDatastore_Task, 'duration_secs': 0.007027} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.635666] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.635910] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.642134] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240050, 'name': Destroy_Task, 'duration_secs': 0.6614} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.642394] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Destroyed the VM [ 1063.642675] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Deleting Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1063.642955] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f38ec06a-6e48-439f-a9a7-1082ff698383 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.649910] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1063.649910] env[62552]: value = "task-1240054" [ 1063.649910] env[62552]: _type = "Task" [ 1063.649910] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.659741] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240054, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.922276] env[62552]: DEBUG oslo_concurrency.lockutils [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "462a3224-d843-45f1-a2bf-69afc18e2572" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.922729] env[62552]: DEBUG oslo_concurrency.lockutils [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.923110] env[62552]: INFO nova.compute.manager [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Attaching volume 90b18b24-f92b-4732-82b5-d40f557a142a to /dev/sdb [ 1063.960285] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3341019-c19a-482f-8d55-cb927112effd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.967699] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b9c41e-a0ed-4f54-b64e-4d4abb8fab8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.980328] env[62552]: DEBUG nova.virt.block_device [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Updating existing volume attachment record: ee6c9bc9-b72c-4061-9e57-d09c81744de1 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1064.006481] env[62552]: INFO nova.compute.manager [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Took 14.18 seconds to build instance. [ 1064.052419] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240051, 'name': ReconfigVM_Task, 'duration_secs': 0.650962} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.053467] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 6824a837-60d0-4577-927f-a54d28a68f80/6824a837-60d0-4577-927f-a54d28a68f80.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.054649] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db99b7c2-bf19-4871-8819-aef01ea1e7a2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.061104] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1064.061104] env[62552]: value = "task-1240055" [ 1064.061104] env[62552]: _type = "Task" [ 1064.061104] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.073969] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240055, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.115695] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240053, 'name': ReconfigVM_Task, 'duration_secs': 0.304723} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.116096] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 7e319ad7-ca42-4eb1-b303-88adbfba7bfe/7e319ad7-ca42-4eb1-b303-88adbfba7bfe.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.116875] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-caefedfa-d4a2-4293-bae1-4417731e8c55 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.122603] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1064.122603] env[62552]: value = "task-1240057" [ 1064.122603] env[62552]: _type = "Task" [ 1064.122603] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.131662] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240057, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.161396] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240054, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.337389] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ff3148-99b3-4ecd-83f8-780ca1a6bbc7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.346351] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ce4980-c530-4286-bb5e-00f5576e75f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.390636] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ba5ff2-0d24-4a30-a9ba-375361cc3a1c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.401255] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8fc4d6-bcac-45aa-b8ab-fc87d9390cda {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.423233] env[62552]: DEBUG nova.compute.provider_tree [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.509670] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e16ea40e-5c6e-4493-89e2-1361c322b034 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "68cb79de-e08c-4ceb-acca-e45952f67248" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.701s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.540901] env[62552]: DEBUG oslo_concurrency.lockutils [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "79166f2c-f864-4d8a-b1dc-e176710400cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.541661] env[62552]: DEBUG oslo_concurrency.lockutils [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.541661] env[62552]: DEBUG oslo_concurrency.lockutils [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "79166f2c-f864-4d8a-b1dc-e176710400cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.541760] env[62552]: DEBUG oslo_concurrency.lockutils [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.542013] env[62552]: DEBUG oslo_concurrency.lockutils [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.545804] env[62552]: INFO nova.compute.manager [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Terminating instance [ 1064.563150] env[62552]: DEBUG nova.compute.manager [req-30d1d88a-38b5-4a52-aaae-5ca0c1b59bd8 req-4d42589f-4a8b-4718-9023-d5283c14c126 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Received event network-changed-4f6cbf66-fc98-4868-82ef-2f234be029cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1064.563150] env[62552]: DEBUG nova.compute.manager [req-30d1d88a-38b5-4a52-aaae-5ca0c1b59bd8 req-4d42589f-4a8b-4718-9023-d5283c14c126 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Refreshing instance network info cache due to event network-changed-4f6cbf66-fc98-4868-82ef-2f234be029cb. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1064.563150] env[62552]: DEBUG oslo_concurrency.lockutils [req-30d1d88a-38b5-4a52-aaae-5ca0c1b59bd8 req-4d42589f-4a8b-4718-9023-d5283c14c126 service nova] Acquiring lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.563150] env[62552]: DEBUG oslo_concurrency.lockutils [req-30d1d88a-38b5-4a52-aaae-5ca0c1b59bd8 req-4d42589f-4a8b-4718-9023-d5283c14c126 service nova] Acquired lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.563150] env[62552]: DEBUG nova.network.neutron [req-30d1d88a-38b5-4a52-aaae-5ca0c1b59bd8 req-4d42589f-4a8b-4718-9023-d5283c14c126 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Refreshing network info cache for port 4f6cbf66-fc98-4868-82ef-2f234be029cb {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1064.577178] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240055, 'name': Rename_Task, 'duration_secs': 0.135256} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.577500] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.577736] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-092915ff-9f2c-46f7-9f98-d28fc1fb7ac4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.585648] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1064.585648] env[62552]: value = "task-1240060" [ 1064.585648] env[62552]: _type = "Task" [ 1064.585648] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.594268] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240060, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.633777] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240057, 'name': Rename_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.660300] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240054, 'name': RemoveSnapshot_Task, 'duration_secs': 0.841419} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.660560] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Deleted Snapshot of the VM instance {{(pid=62552) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1064.660878] env[62552]: DEBUG nova.compute.manager [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1064.661684] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0006f244-d48e-45e1-b7d6-b66d94868ee8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.927056] env[62552]: DEBUG nova.scheduler.client.report [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1065.053234] env[62552]: DEBUG nova.compute.manager [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1065.053487] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1065.054490] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f58123b-54af-4e56-9ada-eb4c39156773 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.062343] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1065.062644] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44d9218c-ed35-4195-8a2d-4f8befcd5a54 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.070440] env[62552]: DEBUG oslo_vmware.api [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1065.070440] env[62552]: value = "task-1240061" [ 1065.070440] env[62552]: _type = "Task" [ 1065.070440] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.079386] env[62552]: DEBUG oslo_vmware.api [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.094521] env[62552]: DEBUG oslo_vmware.api [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240060, 'name': PowerOnVM_Task, 'duration_secs': 0.426953} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.094810] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.095027] env[62552]: INFO nova.compute.manager [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Took 5.88 seconds to spawn the instance on the hypervisor. [ 1065.095220] env[62552]: DEBUG nova.compute.manager [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1065.096037] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc46525-d759-41a5-b9ec-9cf85047207c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.133777] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240057, 'name': Rename_Task, 'duration_secs': 0.85668} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.134289] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.134362] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c664099-246e-4bfa-a522-5eebc0696fe4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.140774] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1065.140774] env[62552]: value = "task-1240062" [ 1065.140774] env[62552]: _type = "Task" [ 1065.140774] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.149568] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240062, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.173726] env[62552]: INFO nova.compute.manager [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Shelve offloading [ 1065.378676] env[62552]: DEBUG nova.network.neutron [req-30d1d88a-38b5-4a52-aaae-5ca0c1b59bd8 req-4d42589f-4a8b-4718-9023-d5283c14c126 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updated VIF entry in instance network info cache for port 4f6cbf66-fc98-4868-82ef-2f234be029cb. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1065.379090] env[62552]: DEBUG nova.network.neutron [req-30d1d88a-38b5-4a52-aaae-5ca0c1b59bd8 req-4d42589f-4a8b-4718-9023-d5283c14c126 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance_info_cache with network_info: [{"id": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "address": "fa:16:3e:9c:12:00", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f6cbf66-fc", "ovs_interfaceid": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.580907] env[62552]: DEBUG oslo_vmware.api [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240061, 'name': PowerOffVM_Task, 'duration_secs': 0.294559} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.581204] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.581379] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1065.581633] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42fad0df-0d62-4ec4-98d8-5d51c8930657 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.617969] env[62552]: INFO nova.compute.manager [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Took 11.48 seconds to build instance. [ 1065.651459] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240062, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.660468] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1065.660656] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1065.660847] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Deleting the datastore file [datastore2] 79166f2c-f864-4d8a-b1dc-e176710400cb {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1065.661429] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34bd5702-dc93-4962-8000-c6d041571edd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.668167] env[62552]: DEBUG oslo_vmware.api [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1065.668167] env[62552]: value = "task-1240064" [ 1065.668167] env[62552]: _type = "Task" [ 1065.668167] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.679030] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1065.679344] env[62552]: DEBUG oslo_vmware.api [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240064, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.679592] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5bc305a-a60c-4da2-b038-622d21a8732c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.685704] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1065.685704] env[62552]: value = "task-1240065" [ 1065.685704] env[62552]: _type = "Task" [ 1065.685704] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.693581] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240065, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.882206] env[62552]: DEBUG oslo_concurrency.lockutils [req-30d1d88a-38b5-4a52-aaae-5ca0c1b59bd8 req-4d42589f-4a8b-4718-9023-d5283c14c126 service nova] Releasing lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.942529] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.306s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.120418] env[62552]: DEBUG oslo_concurrency.lockutils [None req-67052ecf-0b49-44ce-9430-80908cc2129d tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "6824a837-60d0-4577-927f-a54d28a68f80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.998s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.153051] env[62552]: DEBUG oslo_vmware.api [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240062, 'name': PowerOnVM_Task, 'duration_secs': 0.559518} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.153381] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1066.153381] env[62552]: INFO nova.compute.manager [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Took 9.24 seconds to spawn the instance on the hypervisor. [ 1066.153566] env[62552]: DEBUG nova.compute.manager [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1066.154420] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0181d6-4aba-4cd4-810a-91f7da7264c7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.177534] env[62552]: DEBUG oslo_vmware.api [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240064, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216556} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.177843] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1066.178087] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1066.178309] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1066.178520] env[62552]: INFO nova.compute.manager [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1066.178801] env[62552]: DEBUG oslo.service.loopingcall [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.179042] env[62552]: DEBUG nova.compute.manager [-] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1066.179171] env[62552]: DEBUG nova.network.neutron [-] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1066.196148] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1066.196370] env[62552]: DEBUG nova.compute.manager [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1066.197138] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb4359a-64f2-489d-8a20-653769966a12 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.207133] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.207133] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.207133] env[62552]: DEBUG nova.network.neutron [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.360822] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "60238708-4d45-4d73-bd8d-2cbc25704259" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.361108] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.508830] env[62552]: INFO nova.scheduler.client.report [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted allocation for migration 0cdd2c76-f597-4186-8b46-e13da9e1da53 [ 1066.674463] env[62552]: INFO nova.compute.manager [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Took 14.14 seconds to build instance. [ 1066.701748] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Didn't find any instances for network info cache update. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10357}} [ 1066.702146] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.702364] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.702816] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.702816] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.702816] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.702984] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.703110] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1066.703253] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.818710] env[62552]: DEBUG nova.compute.manager [None req-ca639bd2-3905-49e9-8192-5681ecf5b7dc tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1066.819817] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f837f4-6a32-44fc-b5b6-4757a05e6346 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.863696] env[62552]: DEBUG nova.compute.manager [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1067.004550] env[62552]: DEBUG nova.compute.manager [req-53bdbbb6-1fef-43ba-9a95-9a8dbc9c4d83 req-c0a75d78-b6f4-4daf-8142-86956a3d3a06 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Received event network-vif-deleted-19a87259-1f7a-428c-8790-b74e27fde6b6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1067.004791] env[62552]: INFO nova.compute.manager [req-53bdbbb6-1fef-43ba-9a95-9a8dbc9c4d83 req-c0a75d78-b6f4-4daf-8142-86956a3d3a06 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Neutron deleted interface 19a87259-1f7a-428c-8790-b74e27fde6b6; detaching it from the instance and deleting it from the info cache [ 1067.004934] env[62552]: DEBUG nova.network.neutron [req-53bdbbb6-1fef-43ba-9a95-9a8dbc9c4d83 req-c0a75d78-b6f4-4daf-8142-86956a3d3a06 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.015151] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.694s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.035501] env[62552]: DEBUG nova.network.neutron [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updating instance_info_cache with network_info: [{"id": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "address": "fa:16:3e:38:6f:9c", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1028f0-d5", "ovs_interfaceid": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.036764] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquiring lock "6824a837-60d0-4577-927f-a54d28a68f80" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.036997] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "6824a837-60d0-4577-927f-a54d28a68f80" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.037214] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquiring lock "6824a837-60d0-4577-927f-a54d28a68f80-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.037398] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "6824a837-60d0-4577-927f-a54d28a68f80-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.037563] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "6824a837-60d0-4577-927f-a54d28a68f80-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.040803] env[62552]: INFO nova.compute.manager [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Terminating instance [ 1067.179617] env[62552]: DEBUG oslo_concurrency.lockutils [None req-118775f8-0311-4878-82de-2c902d6894d8 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.654s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.181334] env[62552]: DEBUG oslo_concurrency.lockutils [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.181604] env[62552]: DEBUG oslo_concurrency.lockutils [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.181850] env[62552]: DEBUG oslo_concurrency.lockutils [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.182140] env[62552]: DEBUG oslo_concurrency.lockutils [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.182345] env[62552]: DEBUG oslo_concurrency.lockutils [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.184550] env[62552]: INFO nova.compute.manager [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Terminating instance [ 1067.195101] env[62552]: DEBUG nova.network.neutron [-] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.206879] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.207380] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.207380] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.207495] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1067.208511] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f0d7bd-4596-4750-8399-248015975dca {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.217374] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b547eb6-0436-4f8e-965e-184f1813d238 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.231709] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53d47ac-f449-4b40-b540-c9084769bdf7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.238718] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb17502-3826-4102-a5af-9f4ee2f1728d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.270894] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179808MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1067.271059] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.271258] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.330152] env[62552]: INFO nova.compute.manager [None req-ca639bd2-3905-49e9-8192-5681ecf5b7dc tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] instance snapshotting [ 1067.330826] env[62552]: DEBUG nova.objects.instance [None req-ca639bd2-3905-49e9-8192-5681ecf5b7dc tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lazy-loading 'flavor' on Instance uuid 6824a837-60d0-4577-927f-a54d28a68f80 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.388204] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.507444] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-efc51190-134a-47e0-8610-580157d3cbe1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.517476] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67a45bc-fadc-4319-829f-5978beadb471 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.547352] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.549366] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquiring lock "refresh_cache-6824a837-60d0-4577-927f-a54d28a68f80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.549534] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquired lock "refresh_cache-6824a837-60d0-4577-927f-a54d28a68f80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.549705] env[62552]: DEBUG nova.network.neutron [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1067.551098] env[62552]: DEBUG nova.compute.manager [req-53bdbbb6-1fef-43ba-9a95-9a8dbc9c4d83 req-c0a75d78-b6f4-4daf-8142-86956a3d3a06 service nova] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Detach interface failed, port_id=19a87259-1f7a-428c-8790-b74e27fde6b6, reason: Instance 79166f2c-f864-4d8a-b1dc-e176710400cb could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1067.688729] env[62552]: DEBUG nova.compute.manager [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1067.689019] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.690349] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d331b907-a8b9-442c-b6ec-a7abb23c9032 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.698783] env[62552]: INFO nova.compute.manager [-] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Took 1.52 seconds to deallocate network for instance. [ 1067.699021] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.700720] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a8cb724-4f3d-4f69-b2be-d6b11411b403 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.709319] env[62552]: DEBUG oslo_vmware.api [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1067.709319] env[62552]: value = "task-1240067" [ 1067.709319] env[62552]: _type = "Task" [ 1067.709319] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.717584] env[62552]: DEBUG oslo_vmware.api [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240067, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.837199] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1114cf3e-2a3b-4a76-829c-ed8750dfb8c4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.858834] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16d0173-86ed-47b8-8f14-dde373f6f03c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.893121] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.894070] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af82315-9575-41b8-b0d6-2430b696c15e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.902623] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.902911] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97a080a7-dad9-4542-95fd-c3c7799f4eec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.995099] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.995315] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.995501] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleting the datastore file [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.995766] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08e15e9f-90e6-4813-97ba-1cb54b0455c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.002053] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1068.002053] env[62552]: value = "task-1240069" [ 1068.002053] env[62552]: _type = "Task" [ 1068.002053] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.010199] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240069, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.075338] env[62552]: DEBUG nova.network.neutron [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1068.149133] env[62552]: DEBUG nova.network.neutron [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.207466] env[62552]: DEBUG oslo_concurrency.lockutils [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.218615] env[62552]: DEBUG oslo_vmware.api [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240067, 'name': PowerOffVM_Task, 'duration_secs': 0.201913} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.218876] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.219064] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1068.219314] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a77afe8-d3aa-4229-ac96-eb65d6d38a35 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.288026] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1068.288026] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1068.288221] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Deleting the datastore file [datastore2] 7e319ad7-ca42-4eb1-b303-88adbfba7bfe {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.288799] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7012eae8-5ba7-4d69-af89-e9be81a89ca8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.295555] env[62552]: DEBUG oslo_vmware.api [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for the task: (returnval){ [ 1068.295555] env[62552]: value = "task-1240071" [ 1068.295555] env[62552]: _type = "Task" [ 1068.295555] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.303009] env[62552]: DEBUG oslo_vmware.api [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240071, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.303904] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 11831580-1b58-476a-91ce-a4e55947fd91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.304054] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6c8b5367-4fa0-479b-9382-ff261201e3ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.304181] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6a5b5f4e-0c84-447d-a3da-8258512abb7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.304296] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 79166f2c-f864-4d8a-b1dc-e176710400cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.304409] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 158d4227-90ca-41ae-821b-efd353928cb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.304520] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 462a3224-d843-45f1-a2bf-69afc18e2572 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.304632] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 765bbf00-2569-4c49-9634-81d10f375175 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.304742] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance e3f9d5f3-705a-46b1-80d0-bd2421a8b074 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.304850] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 68cb79de-e08c-4ceb-acca-e45952f67248 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.304964] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 7e319ad7-ca42-4eb1-b303-88adbfba7bfe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.305087] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6824a837-60d0-4577-927f-a54d28a68f80 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.346391] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.346654] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.346865] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.347065] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.347243] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.349612] env[62552]: INFO nova.compute.manager [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Terminating instance [ 1068.370505] env[62552]: DEBUG nova.compute.manager [None req-ca639bd2-3905-49e9-8192-5681ecf5b7dc tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Instance disappeared during snapshot {{(pid=62552) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 1068.494630] env[62552]: DEBUG nova.compute.manager [None req-ca639bd2-3905-49e9-8192-5681ecf5b7dc tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Found 0 images (rotation: 2) {{(pid=62552) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4897}} [ 1068.513637] env[62552]: DEBUG oslo_vmware.api [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240069, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130802} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.514660] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.514920] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.515692] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.526299] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1068.526624] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1068.528361] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed10518e-50da-4ff3-8368-88480a8dde0c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.549244] env[62552]: INFO nova.scheduler.client.report [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted allocations for instance 158d4227-90ca-41ae-821b-efd353928cb2 [ 1068.552749] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9deca59c-9b0d-4834-96a3-15f064e8260d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.579100] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-90b18b24-f92b-4732-82b5-d40f557a142a/volume-90b18b24-f92b-4732-82b5-d40f557a142a.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.579421] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4ba4ca1-7a07-4ca1-905b-b7ddeace2bea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.601390] env[62552]: DEBUG oslo_vmware.api [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1068.601390] env[62552]: value = "task-1240072" [ 1068.601390] env[62552]: _type = "Task" [ 1068.601390] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.610398] env[62552]: DEBUG oslo_vmware.api [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240072, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.652037] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Releasing lock "refresh_cache-6824a837-60d0-4577-927f-a54d28a68f80" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.652544] env[62552]: DEBUG nova.compute.manager [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1068.652831] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1068.653720] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a4ab60-3bb5-4971-b845-fbfa24b70023 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.661340] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.661599] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f5d1cd7-af89-42a2-bb8f-41ff0ecd1982 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.668335] env[62552]: DEBUG oslo_vmware.api [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1068.668335] env[62552]: value = "task-1240073" [ 1068.668335] env[62552]: _type = "Task" [ 1068.668335] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.679771] env[62552]: DEBUG oslo_vmware.api [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240073, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.805430] env[62552]: DEBUG oslo_vmware.api [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Task: {'id': task-1240071, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141471} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.805650] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.805825] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.806016] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.806198] env[62552]: INFO nova.compute.manager [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1068.806438] env[62552]: DEBUG oslo.service.loopingcall [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.806631] env[62552]: DEBUG nova.compute.manager [-] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1068.806727] env[62552]: DEBUG nova.network.neutron [-] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1068.808975] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 60238708-4d45-4d73-bd8d-2cbc25704259 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1068.809219] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1068.809361] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1068.852910] env[62552]: DEBUG nova.compute.manager [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1068.853153] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1068.854411] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab67e52d-53a8-4da5-81a3-bf67dc2b2e6c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.865562] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.865904] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5e93055-85b0-48ce-b5b9-5789d0a6dc66 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.872499] env[62552]: DEBUG oslo_vmware.api [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1068.872499] env[62552]: value = "task-1240074" [ 1068.872499] env[62552]: _type = "Task" [ 1068.872499] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.881788] env[62552]: DEBUG oslo_vmware.api [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240074, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.958706] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4a3750-e57d-40ef-b35f-830b52f6b757 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.969262] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8855ce9-d535-4d53-8fc5-97dfb83b343b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.001709] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057b514f-1dea-49e1-ae5c-425b5706cb73 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.010110] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30bda032-2614-4db5-941f-1d58b6f522ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.025235] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.031062] env[62552]: DEBUG nova.compute.manager [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received event network-vif-unplugged-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1069.031286] env[62552]: DEBUG oslo_concurrency.lockutils [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] Acquiring lock "158d4227-90ca-41ae-821b-efd353928cb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.031496] env[62552]: DEBUG oslo_concurrency.lockutils [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] Lock "158d4227-90ca-41ae-821b-efd353928cb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.031669] env[62552]: DEBUG oslo_concurrency.lockutils [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] Lock "158d4227-90ca-41ae-821b-efd353928cb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.031889] env[62552]: DEBUG nova.compute.manager [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] No waiting events found dispatching network-vif-unplugged-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1069.031990] env[62552]: WARNING nova.compute.manager [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received unexpected event network-vif-unplugged-dd1028f0-d5a9-48be-84e3-9401cde0a1ff for instance with vm_state shelved_offloaded and task_state None. [ 1069.032159] env[62552]: DEBUG nova.compute.manager [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received event network-changed-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1069.032313] env[62552]: DEBUG nova.compute.manager [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Refreshing instance network info cache due to event network-changed-dd1028f0-d5a9-48be-84e3-9401cde0a1ff. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1069.033116] env[62552]: DEBUG oslo_concurrency.lockutils [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] Acquiring lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.033116] env[62552]: DEBUG oslo_concurrency.lockutils [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] Acquired lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.033116] env[62552]: DEBUG nova.network.neutron [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Refreshing network info cache for port dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1069.056839] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.111045] env[62552]: DEBUG oslo_vmware.api [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240072, 'name': ReconfigVM_Task, 'duration_secs': 0.409892} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.111380] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-90b18b24-f92b-4732-82b5-d40f557a142a/volume-90b18b24-f92b-4732-82b5-d40f557a142a.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.116037] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bc7d7cf-e423-42c3-8f1e-96cc1231f8a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.130532] env[62552]: DEBUG oslo_vmware.api [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1069.130532] env[62552]: value = "task-1240075" [ 1069.130532] env[62552]: _type = "Task" [ 1069.130532] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.138517] env[62552]: DEBUG oslo_vmware.api [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240075, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.177333] env[62552]: DEBUG oslo_vmware.api [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240073, 'name': PowerOffVM_Task, 'duration_secs': 0.100585} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.177600] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1069.177767] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.178014] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb7f2ce0-90b8-426b-b42f-24f9a30be3b4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.201827] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.202358] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.202358] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Deleting the datastore file [datastore2] 6824a837-60d0-4577-927f-a54d28a68f80 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.202514] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae21814c-d16b-4d1c-a713-9556aabf76f8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.208744] env[62552]: DEBUG oslo_vmware.api [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for the task: (returnval){ [ 1069.208744] env[62552]: value = "task-1240077" [ 1069.208744] env[62552]: _type = "Task" [ 1069.208744] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.216169] env[62552]: DEBUG oslo_vmware.api [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240077, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.381566] env[62552]: DEBUG oslo_vmware.api [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240074, 'name': PowerOffVM_Task, 'duration_secs': 0.225377} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.381847] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1069.382029] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.382309] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-65feab23-9a11-434c-92f9-db28f70f09ac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.459879] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.460199] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.460395] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleting the datastore file [datastore1] e3f9d5f3-705a-46b1-80d0-bd2421a8b074 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.460681] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf01186d-45cb-4f04-8ce9-2afbc3aa507c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.467294] env[62552]: DEBUG oslo_vmware.api [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for the task: (returnval){ [ 1069.467294] env[62552]: value = "task-1240079" [ 1069.467294] env[62552]: _type = "Task" [ 1069.467294] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.475128] env[62552]: DEBUG oslo_vmware.api [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240079, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.528376] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1069.592624] env[62552]: DEBUG nova.network.neutron [-] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.644036] env[62552]: DEBUG oslo_vmware.api [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240075, 'name': ReconfigVM_Task, 'duration_secs': 0.139546} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.644375] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1069.720191] env[62552]: DEBUG oslo_vmware.api [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Task: {'id': task-1240077, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089192} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.721023] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.721023] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.721281] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.721641] env[62552]: INFO nova.compute.manager [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1069.722277] env[62552]: DEBUG oslo.service.loopingcall [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.722416] env[62552]: DEBUG nova.compute.manager [-] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1069.722767] env[62552]: DEBUG nova.network.neutron [-] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1069.742172] env[62552]: DEBUG nova.network.neutron [-] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1069.772397] env[62552]: DEBUG nova.network.neutron [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updated VIF entry in instance network info cache for port dd1028f0-d5a9-48be-84e3-9401cde0a1ff. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1069.773108] env[62552]: DEBUG nova.network.neutron [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updating instance_info_cache with network_info: [{"id": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "address": "fa:16:3e:38:6f:9c", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapdd1028f0-d5", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.977710] env[62552]: DEBUG oslo_vmware.api [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Task: {'id': task-1240079, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138847} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.979047] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.979047] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.979047] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.979047] env[62552]: INFO nova.compute.manager [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1069.979047] env[62552]: DEBUG oslo.service.loopingcall [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.979288] env[62552]: DEBUG nova.compute.manager [-] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1069.979333] env[62552]: DEBUG nova.network.neutron [-] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1070.033013] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1070.033232] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.762s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.033516] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.645s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.035120] env[62552]: INFO nova.compute.claims [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1070.097339] env[62552]: INFO nova.compute.manager [-] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Took 1.29 seconds to deallocate network for instance. [ 1070.244029] env[62552]: DEBUG nova.network.neutron [-] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.275943] env[62552]: DEBUG oslo_concurrency.lockutils [req-63ce0804-7311-4c32-9119-9b545cb63596 req-9566741c-6314-44dd-be88-a29f03039eb8 service nova] Releasing lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.604893] env[62552]: DEBUG oslo_concurrency.lockutils [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.685691] env[62552]: DEBUG nova.objects.instance [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lazy-loading 'flavor' on Instance uuid 462a3224-d843-45f1-a2bf-69afc18e2572 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.693992] env[62552]: DEBUG nova.network.neutron [-] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.747024] env[62552]: INFO nova.compute.manager [-] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Took 1.02 seconds to deallocate network for instance. [ 1071.065018] env[62552]: DEBUG nova.compute.manager [req-dc48f246-fd23-4ed6-b0c0-61df991fc323 req-3cef7665-50b1-4dd7-ad5d-c938c1ed53c5 service nova] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Received event network-vif-deleted-f74cd14f-2def-457c-b333-023c8f6d6567 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1071.065277] env[62552]: DEBUG nova.compute.manager [req-dc48f246-fd23-4ed6-b0c0-61df991fc323 req-3cef7665-50b1-4dd7-ad5d-c938c1ed53c5 service nova] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Received event network-vif-deleted-69937c5b-d6a4-4352-a6b0-ec0ce9d6b80f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1071.190262] env[62552]: DEBUG oslo_concurrency.lockutils [None req-183eaf0a-9c49-4848-b355-6bdc51985802 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.267s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.198835] env[62552]: INFO nova.compute.manager [-] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Took 1.22 seconds to deallocate network for instance. [ 1071.218183] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d62e74-ccac-481d-bcaa-fd9c6a2b295e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.226866] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d570219-6638-405a-aed2-c845ea963953 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.261943] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.263031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "158d4227-90ca-41ae-821b-efd353928cb2" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.264284] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14101d41-112b-452f-b962-b7fc7144f99f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.272856] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c598ff9e-6a71-4832-89e0-8fae37a39001 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.286373] env[62552]: DEBUG nova.compute.provider_tree [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.577589] env[62552]: INFO nova.compute.manager [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Rebuilding instance [ 1071.612783] env[62552]: DEBUG nova.compute.manager [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1071.613654] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e0b972-41c7-4494-9ca3-b8e55d3bd2dc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.709180] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.788681] env[62552]: DEBUG nova.scheduler.client.report [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1072.295036] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.295036] env[62552]: DEBUG nova.compute.manager [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1072.297032] env[62552]: DEBUG oslo_concurrency.lockutils [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.090s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.297252] env[62552]: DEBUG nova.objects.instance [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lazy-loading 'resources' on Instance uuid 79166f2c-f864-4d8a-b1dc-e176710400cb {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.627575] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.627906] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5cbe834b-1be9-4ba9-859e-e69222a664d5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.637894] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1072.637894] env[62552]: value = "task-1240080" [ 1072.637894] env[62552]: _type = "Task" [ 1072.637894] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.646309] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240080, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.800586] env[62552]: DEBUG nova.compute.utils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1072.805381] env[62552]: DEBUG nova.compute.manager [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1072.805560] env[62552]: DEBUG nova.network.neutron [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1072.877033] env[62552]: DEBUG nova.policy [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecfe0d4d977540d99709760db0564cf9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3758acde6e3e4320bf63d4cd1667bfc2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1072.948403] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2e5460-6dc0-4f05-b874-5fb1865dfb98 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.956773] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6027dd7-0ac5-490c-aa23-4830752d6f77 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.986970] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683d74c3-7c14-4273-ba58-973701ed7aae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.994393] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a82ff5-2acc-4b04-a295-5ea36e5e0c6e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.007417] env[62552]: DEBUG nova.compute.provider_tree [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.148208] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240080, 'name': PowerOffVM_Task, 'duration_secs': 0.224483} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.148545] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1073.177283] env[62552]: DEBUG nova.network.neutron [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Successfully created port: a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1073.212134] env[62552]: INFO nova.compute.manager [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Detaching volume 90b18b24-f92b-4732-82b5-d40f557a142a [ 1073.243340] env[62552]: INFO nova.virt.block_device [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Attempting to driver detach volume 90b18b24-f92b-4732-82b5-d40f557a142a from mountpoint /dev/sdb [ 1073.243549] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1073.243743] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1073.244669] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51aa3497-e18e-4d14-8ce6-da947cf9d79d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.268191] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db361cd0-1757-430e-a330-4b0bf4fea32f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.275475] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8c50d5-f843-403d-85fd-07c655074fce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.295578] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b56890-be69-4978-9b77-efa7014b41b6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.311468] env[62552]: DEBUG nova.compute.manager [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1073.314228] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] The volume has not been displaced from its original location: [datastore1] volume-90b18b24-f92b-4732-82b5-d40f557a142a/volume-90b18b24-f92b-4732-82b5-d40f557a142a.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1073.319421] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1073.320217] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9851264e-a143-47a8-a543-5219cdbd916a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.339684] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1073.339684] env[62552]: value = "task-1240081" [ 1073.339684] env[62552]: _type = "Task" [ 1073.339684] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.348506] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.510914] env[62552]: DEBUG nova.scheduler.client.report [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1073.850631] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240081, 'name': ReconfigVM_Task, 'duration_secs': 0.193819} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.850631] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1073.855113] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9d737bc-7b2a-40c2-9a34-56fca7578e50 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.871261] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1073.871261] env[62552]: value = "task-1240082" [ 1073.871261] env[62552]: _type = "Task" [ 1073.871261] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.880527] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240082, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.017063] env[62552]: DEBUG oslo_concurrency.lockutils [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.018936] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.962s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.019739] env[62552]: DEBUG nova.objects.instance [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'resources' on Instance uuid 158d4227-90ca-41ae-821b-efd353928cb2 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.048017] env[62552]: INFO nova.scheduler.client.report [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Deleted allocations for instance 79166f2c-f864-4d8a-b1dc-e176710400cb [ 1074.327409] env[62552]: DEBUG nova.compute.manager [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1074.351682] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1074.351936] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1074.352112] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1074.352322] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1074.352495] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1074.352649] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1074.352924] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1074.353106] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1074.353278] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1074.353443] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1074.353615] env[62552]: DEBUG nova.virt.hardware [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1074.354848] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c04339-c096-44c5-81d9-10353877517a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.364557] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba79f915-447d-46b2-904d-5d518827d2b4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.384439] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240082, 'name': ReconfigVM_Task, 'duration_secs': 0.160482} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.384698] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1074.521797] env[62552]: DEBUG nova.objects.instance [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'numa_topology' on Instance uuid 158d4227-90ca-41ae-821b-efd353928cb2 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.557070] env[62552]: DEBUG oslo_concurrency.lockutils [None req-563a3590-d6e8-422d-ab1c-68fee39a07f0 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "79166f2c-f864-4d8a-b1dc-e176710400cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.016s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.657866] env[62552]: DEBUG nova.compute.manager [req-5a6514de-e377-4d4c-8e5f-0851fcc484cd req-d0dc3106-508f-43d0-8ed2-bab6c9aefd0f service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Received event network-vif-plugged-a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1074.658113] env[62552]: DEBUG oslo_concurrency.lockutils [req-5a6514de-e377-4d4c-8e5f-0851fcc484cd req-d0dc3106-508f-43d0-8ed2-bab6c9aefd0f service nova] Acquiring lock "60238708-4d45-4d73-bd8d-2cbc25704259-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.658335] env[62552]: DEBUG oslo_concurrency.lockutils [req-5a6514de-e377-4d4c-8e5f-0851fcc484cd req-d0dc3106-508f-43d0-8ed2-bab6c9aefd0f service nova] Lock "60238708-4d45-4d73-bd8d-2cbc25704259-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.658491] env[62552]: DEBUG oslo_concurrency.lockutils [req-5a6514de-e377-4d4c-8e5f-0851fcc484cd req-d0dc3106-508f-43d0-8ed2-bab6c9aefd0f service nova] Lock "60238708-4d45-4d73-bd8d-2cbc25704259-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.658662] env[62552]: DEBUG nova.compute.manager [req-5a6514de-e377-4d4c-8e5f-0851fcc484cd req-d0dc3106-508f-43d0-8ed2-bab6c9aefd0f service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] No waiting events found dispatching network-vif-plugged-a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1074.658831] env[62552]: WARNING nova.compute.manager [req-5a6514de-e377-4d4c-8e5f-0851fcc484cd req-d0dc3106-508f-43d0-8ed2-bab6c9aefd0f service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Received unexpected event network-vif-plugged-a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 for instance with vm_state building and task_state spawning. [ 1074.780221] env[62552]: DEBUG nova.network.neutron [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Successfully updated port: a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1075.024622] env[62552]: DEBUG nova.objects.base [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Object Instance<158d4227-90ca-41ae-821b-efd353928cb2> lazy-loaded attributes: resources,numa_topology {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1075.163573] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8626936-adb9-4e09-be13-8c4a7f96fb4f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.172110] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47310bf9-9b0f-4658-96d8-a05fdb99d596 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.201935] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedde86c-9cab-40aa-9064-b807d76ccc57 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.210325] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad0a8cc-5f62-4067-b037-78fb4ca18190 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.225163] env[62552]: DEBUG nova.compute.provider_tree [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.283827] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "refresh_cache-60238708-4d45-4d73-bd8d-2cbc25704259" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.284032] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "refresh_cache-60238708-4d45-4d73-bd8d-2cbc25704259" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.284203] env[62552]: DEBUG nova.network.neutron [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1075.435511] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1075.435844] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59b4323c-b9c5-45b2-b8cf-81c2710934a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.443961] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1075.443961] env[62552]: value = "task-1240083" [ 1075.443961] env[62552]: _type = "Task" [ 1075.443961] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.452179] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240083, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.728162] env[62552]: DEBUG nova.scheduler.client.report [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1075.824526] env[62552]: DEBUG nova.network.neutron [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1075.954294] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1075.954524] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1075.954722] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1075.955481] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c90a298-72ba-4af7-b0bb-8a8481597dc6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.976503] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd74434-4135-459f-b936-e1a0402e010e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.983147] env[62552]: WARNING nova.virt.vmwareapi.driver [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1075.983380] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1075.984129] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab454d8-e58f-4240-8662-3df10af8be57 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.990237] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1075.990456] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a9eafe1-88f9-4681-88e6-3f6669c3fb3f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.011846] env[62552]: DEBUG nova.network.neutron [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Updating instance_info_cache with network_info: [{"id": "a1d4afdd-37b2-4b59-b447-0ca443e4c7f6", "address": "fa:16:3e:9f:ce:e5", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1d4afdd-37", "ovs_interfaceid": "a1d4afdd-37b2-4b59-b447-0ca443e4c7f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.059694] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1076.059938] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1076.060151] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleting the datastore file [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1076.060420] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b95c800-7414-4b6a-8450-404779f0f953 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.068302] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1076.068302] env[62552]: value = "task-1240086" [ 1076.068302] env[62552]: _type = "Task" [ 1076.068302] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.076273] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240086, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.233417] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.214s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.236030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.631s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.236928] env[62552]: DEBUG nova.objects.instance [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lazy-loading 'resources' on Instance uuid 7e319ad7-ca42-4eb1-b303-88adbfba7bfe {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.514057] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "refresh_cache-60238708-4d45-4d73-bd8d-2cbc25704259" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.514397] env[62552]: DEBUG nova.compute.manager [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Instance network_info: |[{"id": "a1d4afdd-37b2-4b59-b447-0ca443e4c7f6", "address": "fa:16:3e:9f:ce:e5", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1d4afdd-37", "ovs_interfaceid": "a1d4afdd-37b2-4b59-b447-0ca443e4c7f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1076.514847] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:ce:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1d4afdd-37b2-4b59-b447-0ca443e4c7f6', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.523017] env[62552]: DEBUG oslo.service.loopingcall [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.523017] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1076.523017] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4508aa3c-b10f-4e50-91b2-44dc510973ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.544611] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.544611] env[62552]: value = "task-1240087" [ 1076.544611] env[62552]: _type = "Task" [ 1076.544611] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.554172] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240087, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.578076] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240086, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134866} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.578076] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1076.578295] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1076.578381] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1076.684328] env[62552]: DEBUG nova.compute.manager [req-b6c1fbb9-01c7-4259-8101-7b37a5e22856 req-a9b17a49-df8d-4d54-80ec-3d441ae3693d service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Received event network-changed-a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1076.684454] env[62552]: DEBUG nova.compute.manager [req-b6c1fbb9-01c7-4259-8101-7b37a5e22856 req-a9b17a49-df8d-4d54-80ec-3d441ae3693d service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Refreshing instance network info cache due to event network-changed-a1d4afdd-37b2-4b59-b447-0ca443e4c7f6. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1076.684718] env[62552]: DEBUG oslo_concurrency.lockutils [req-b6c1fbb9-01c7-4259-8101-7b37a5e22856 req-a9b17a49-df8d-4d54-80ec-3d441ae3693d service nova] Acquiring lock "refresh_cache-60238708-4d45-4d73-bd8d-2cbc25704259" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.684848] env[62552]: DEBUG oslo_concurrency.lockutils [req-b6c1fbb9-01c7-4259-8101-7b37a5e22856 req-a9b17a49-df8d-4d54-80ec-3d441ae3693d service nova] Acquired lock "refresh_cache-60238708-4d45-4d73-bd8d-2cbc25704259" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.685171] env[62552]: DEBUG nova.network.neutron [req-b6c1fbb9-01c7-4259-8101-7b37a5e22856 req-a9b17a49-df8d-4d54-80ec-3d441ae3693d service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Refreshing network info cache for port a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1076.745985] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1ef1ea85-346d-42c8-8a9b-26fbc228b4e8 tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 29.630s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.747836] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 5.485s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.748296] env[62552]: INFO nova.compute.manager [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Unshelving [ 1076.895955] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9e9125-ed55-469d-a573-8dccba3ef547 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.903943] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea17d5e-c578-4bbb-8c9a-41242a9a0618 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.933815] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ac7605-993c-4dfb-8dd1-165df7793d7a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.941198] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23f2eff-ac95-4414-873b-df68e6752c74 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.954052] env[62552]: DEBUG nova.compute.provider_tree [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1077.057018] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240087, 'name': CreateVM_Task, 'duration_secs': 0.32849} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.057219] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1077.057938] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.058161] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.058728] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1077.058728] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-274d1379-73af-4424-b61d-917cba221550 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.063676] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1077.063676] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a656c2-5aa0-5867-915f-6aa8ed0acb10" [ 1077.063676] env[62552]: _type = "Task" [ 1077.063676] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.071400] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a656c2-5aa0-5867-915f-6aa8ed0acb10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.083044] env[62552]: INFO nova.virt.block_device [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Booting with volume 90b18b24-f92b-4732-82b5-d40f557a142a at /dev/sdb [ 1077.111926] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3bad6b08-5150-4a87-9a84-677f70d4ae3d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.121328] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69c2e98-e5c9-4f49-ae32-66dc44e5e949 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.149140] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb5e89f0-b0dd-4cba-979e-8602fd47318b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.156946] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3332295-a6e6-45dc-a146-8d38921b9af8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.183308] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b75ad82-97a2-4963-8b5a-9938335a6689 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.190866] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c807e18f-eeb2-4161-8b72-cfc64282ce8c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.205602] env[62552]: DEBUG nova.virt.block_device [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Updating existing volume attachment record: 20e0818e-3333-422d-95f3-706dd8f89947 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1077.444921] env[62552]: DEBUG nova.network.neutron [req-b6c1fbb9-01c7-4259-8101-7b37a5e22856 req-a9b17a49-df8d-4d54-80ec-3d441ae3693d service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Updated VIF entry in instance network info cache for port a1d4afdd-37b2-4b59-b447-0ca443e4c7f6. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1077.445332] env[62552]: DEBUG nova.network.neutron [req-b6c1fbb9-01c7-4259-8101-7b37a5e22856 req-a9b17a49-df8d-4d54-80ec-3d441ae3693d service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Updating instance_info_cache with network_info: [{"id": "a1d4afdd-37b2-4b59-b447-0ca443e4c7f6", "address": "fa:16:3e:9f:ce:e5", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1d4afdd-37", "ovs_interfaceid": "a1d4afdd-37b2-4b59-b447-0ca443e4c7f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.473932] env[62552]: ERROR nova.scheduler.client.report [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] [req-e006a026-8a9f-4e31-be06-8125e22ad21b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b52e1b51-8a99-4529-bad8-00cd27fc0fb8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e006a026-8a9f-4e31-be06-8125e22ad21b"}]} [ 1077.490030] env[62552]: DEBUG nova.scheduler.client.report [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1077.503160] env[62552]: DEBUG nova.scheduler.client.report [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1077.503395] env[62552]: DEBUG nova.compute.provider_tree [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1077.514942] env[62552]: DEBUG nova.scheduler.client.report [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1077.533472] env[62552]: DEBUG nova.scheduler.client.report [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1077.576050] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a656c2-5aa0-5867-915f-6aa8ed0acb10, 'name': SearchDatastore_Task, 'duration_secs': 0.01004} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.578770] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.579026] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.579276] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.579419] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.579751] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.580062] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-944e88d2-4afa-4181-ada3-ae01d27c3560 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.589186] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.589380] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1077.592258] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6d73fc5-a776-4362-91b2-99e890334894 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.597970] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1077.597970] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520fd449-1bdd-2080-5df5-a0d26b1858cd" [ 1077.597970] env[62552]: _type = "Task" [ 1077.597970] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.609850] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520fd449-1bdd-2080-5df5-a0d26b1858cd, 'name': SearchDatastore_Task, 'duration_secs': 0.008551} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.610608] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61b934f7-2b77-497e-8d36-308559f9595f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.615908] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1077.615908] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52544671-a5b2-5af2-304c-78d20d5823b2" [ 1077.615908] env[62552]: _type = "Task" [ 1077.615908] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.624094] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52544671-a5b2-5af2-304c-78d20d5823b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.686820] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7933d07b-007b-4e76-a34c-081972c05415 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.694659] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6e377e-7ca3-4772-b77a-8d9afec181a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.725244] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42283ca6-e85b-4732-a9a7-e9da976f7837 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.732341] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3a2872-814e-4501-9233-8fa60149d863 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.745516] env[62552]: DEBUG nova.compute.provider_tree [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1077.768782] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.948573] env[62552]: DEBUG oslo_concurrency.lockutils [req-b6c1fbb9-01c7-4259-8101-7b37a5e22856 req-a9b17a49-df8d-4d54-80ec-3d441ae3693d service nova] Releasing lock "refresh_cache-60238708-4d45-4d73-bd8d-2cbc25704259" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.129546] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52544671-a5b2-5af2-304c-78d20d5823b2, 'name': SearchDatastore_Task, 'duration_secs': 0.008661} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.129941] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.130239] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 60238708-4d45-4d73-bd8d-2cbc25704259/60238708-4d45-4d73-bd8d-2cbc25704259.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1078.130510] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-20b8c622-f87f-466a-84ef-caed433dc2a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.138133] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1078.138133] env[62552]: value = "task-1240088" [ 1078.138133] env[62552]: _type = "Task" [ 1078.138133] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.146392] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.280410] env[62552]: DEBUG nova.scheduler.client.report [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updated inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with generation 136 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1078.280848] env[62552]: DEBUG nova.compute.provider_tree [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 generation from 136 to 137 during operation: update_inventory {{(pid=62552) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1078.281128] env[62552]: DEBUG nova.compute.provider_tree [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1078.650100] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435371} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.650394] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] 60238708-4d45-4d73-bd8d-2cbc25704259/60238708-4d45-4d73-bd8d-2cbc25704259.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1078.650665] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1078.650914] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b785582a-0050-413a-8fef-70f262401ec1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.657970] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1078.657970] env[62552]: value = "task-1240089" [ 1078.657970] env[62552]: _type = "Task" [ 1078.657970] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.666593] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240089, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.786703] env[62552]: DEBUG oslo_concurrency.lockutils [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.551s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.789369] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.528s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.789618] env[62552]: DEBUG nova.objects.instance [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lazy-loading 'resources' on Instance uuid 6824a837-60d0-4577-927f-a54d28a68f80 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.808067] env[62552]: INFO nova.scheduler.client.report [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Deleted allocations for instance 7e319ad7-ca42-4eb1-b303-88adbfba7bfe [ 1079.167711] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240089, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06542} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.168017] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1079.168804] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9335c7a-e70d-4bd3-aaa4-bd320fc8e7f7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.190186] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 60238708-4d45-4d73-bd8d-2cbc25704259/60238708-4d45-4d73-bd8d-2cbc25704259.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.190452] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b1c8ad4-fb9d-4905-967b-2ed3344dedc5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.214385] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1079.214385] env[62552]: value = "task-1240090" [ 1079.214385] env[62552]: _type = "Task" [ 1079.214385] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.223822] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240090, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.314434] env[62552]: DEBUG oslo_concurrency.lockutils [None req-303047f8-c48e-4b75-9630-03266dc361f3 tempest-ImagesOneServerNegativeTestJSON-1587902904 tempest-ImagesOneServerNegativeTestJSON-1587902904-project-member] Lock "7e319ad7-ca42-4eb1-b303-88adbfba7bfe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.133s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.326288] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1079.327102] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1079.327102] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1079.327102] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1079.327102] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1079.327412] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1079.327489] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1079.327638] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1079.327817] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1079.327984] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1079.328187] env[62552]: DEBUG nova.virt.hardware [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1079.329037] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2397469-463c-4415-a50c-a8ae400d6efb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.341471] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00206340-9aa9-40b7-8f06-d54c99f53433 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.357140] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:f9:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81435491-9993-42c8-80d8-8c4f23693ce4', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1079.364468] env[62552]: DEBUG oslo.service.loopingcall [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.367511] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1079.367511] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b31d3bf5-9795-47a3-86f4-152eec2c0b3a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.390507] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1079.390507] env[62552]: value = "task-1240091" [ 1079.390507] env[62552]: _type = "Task" [ 1079.390507] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.401280] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240091, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.457430] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3090eb21-0cf6-4cd1-b666-9b999ec41f1e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.466380] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad197a2e-411b-4dac-b2de-422ee0152623 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.498203] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535b5e6b-66b1-4e7d-9ec5-f5f9c7e32584 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.506582] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5de64b-343d-4c0a-a706-83e0f01a7676 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.522933] env[62552]: DEBUG nova.compute.provider_tree [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.726864] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240090, 'name': ReconfigVM_Task, 'duration_secs': 0.350323} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.730290] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 60238708-4d45-4d73-bd8d-2cbc25704259/60238708-4d45-4d73-bd8d-2cbc25704259.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.730290] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-702a201e-686c-4c02-a169-d156b4bef0a4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.735953] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1079.735953] env[62552]: value = "task-1240092" [ 1079.735953] env[62552]: _type = "Task" [ 1079.735953] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.745581] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240092, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.900875] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240091, 'name': CreateVM_Task, 'duration_secs': 0.367596} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.901338] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1079.902055] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.902249] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.902595] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1079.902895] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92c6bcb2-2571-42a5-a9ec-f46203d5f477 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.907566] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1079.907566] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527ea67b-8e03-7649-2029-cd01dac5d9a1" [ 1079.907566] env[62552]: _type = "Task" [ 1079.907566] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.915641] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527ea67b-8e03-7649-2029-cd01dac5d9a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.027313] env[62552]: DEBUG nova.scheduler.client.report [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1080.216604] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "6c8b5367-4fa0-479b-9382-ff261201e3ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.216871] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "6c8b5367-4fa0-479b-9382-ff261201e3ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.217097] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "6c8b5367-4fa0-479b-9382-ff261201e3ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.217349] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "6c8b5367-4fa0-479b-9382-ff261201e3ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.217529] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "6c8b5367-4fa0-479b-9382-ff261201e3ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.219716] env[62552]: INFO nova.compute.manager [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Terminating instance [ 1080.247707] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240092, 'name': Rename_Task, 'duration_secs': 0.160545} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.248045] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.248345] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f664dc9-2e58-43e6-b6ac-232c38d8fa7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.256792] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1080.256792] env[62552]: value = "task-1240093" [ 1080.256792] env[62552]: _type = "Task" [ 1080.256792] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.263841] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.418151] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527ea67b-8e03-7649-2029-cd01dac5d9a1, 'name': SearchDatastore_Task, 'duration_secs': 0.010305} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.418436] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.418676] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1080.418909] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.419073] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.419260] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1080.419525] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39b1492b-a0ff-49c7-b6a3-c38753dd0d2b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.429802] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1080.429802] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1080.429802] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb19965e-47f9-4c8a-a1b6-568abbe54610 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.437659] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1080.437659] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52675a03-64ec-52d7-732d-c7288a7b890e" [ 1080.437659] env[62552]: _type = "Task" [ 1080.437659] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.443598] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52675a03-64ec-52d7-732d-c7288a7b890e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.533936] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.745s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.536210] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.827s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.536441] env[62552]: DEBUG nova.objects.instance [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lazy-loading 'resources' on Instance uuid e3f9d5f3-705a-46b1-80d0-bd2421a8b074 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.564037] env[62552]: INFO nova.scheduler.client.report [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Deleted allocations for instance 6824a837-60d0-4577-927f-a54d28a68f80 [ 1080.724671] env[62552]: DEBUG nova.compute.manager [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1080.724671] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1080.725882] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54930053-1913-42a3-a0b0-d37634a2b8b0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.734196] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.734196] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca90b3c8-30e2-432a-afa1-19a7d7d404fe {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.740917] env[62552]: DEBUG oslo_vmware.api [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1080.740917] env[62552]: value = "task-1240094" [ 1080.740917] env[62552]: _type = "Task" [ 1080.740917] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.750484] env[62552]: DEBUG oslo_vmware.api [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240094, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.764706] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240093, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.951705] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52675a03-64ec-52d7-732d-c7288a7b890e, 'name': SearchDatastore_Task, 'duration_secs': 0.009788} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.952623] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be7ba75d-c37b-4f6c-a400-6fe3a0a64fcf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.958766] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1080.958766] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52287148-1c26-2f94-413c-87550ee712d8" [ 1080.958766] env[62552]: _type = "Task" [ 1080.958766] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.967733] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52287148-1c26-2f94-413c-87550ee712d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.071571] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c17b9d4c-bba8-4cd1-bbc9-b0a1f0b4e470 tempest-ServersAaction247Test-1086536068 tempest-ServersAaction247Test-1086536068-project-member] Lock "6824a837-60d0-4577-927f-a54d28a68f80" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.034s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.172514] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375ac18b-c87d-4cdc-bcfb-3af4cfad7c8f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.180790] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9ff1e7-5f94-4641-9cb7-1a4a77259347 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.213837] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a750a41-48e2-4d3f-9b33-39edcb455b7b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.222670] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef22cae-8f49-493b-934b-a4f236b1cd33 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.236643] env[62552]: DEBUG nova.compute.provider_tree [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.250308] env[62552]: DEBUG oslo_vmware.api [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240094, 'name': PowerOffVM_Task, 'duration_secs': 0.249947} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.250584] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1081.250783] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1081.251065] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9c07fa9-1483-4c6d-a8af-8bfd2a438677 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.266641] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240093, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.332320] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1081.332623] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1081.332906] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Deleting the datastore file [datastore1] 6c8b5367-4fa0-479b-9382-ff261201e3ef {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1081.333239] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38c10923-5ca4-464b-8011-1bfe0b37e9b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.340946] env[62552]: DEBUG oslo_vmware.api [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1081.340946] env[62552]: value = "task-1240096" [ 1081.340946] env[62552]: _type = "Task" [ 1081.340946] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.349108] env[62552]: DEBUG oslo_vmware.api [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240096, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.470821] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52287148-1c26-2f94-413c-87550ee712d8, 'name': SearchDatastore_Task, 'duration_secs': 0.011407} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.471137] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.471413] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572/462a3224-d843-45f1-a2bf-69afc18e2572.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1081.471692] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23cda373-80ac-40b1-b8b4-abcc16b29e84 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.479839] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1081.479839] env[62552]: value = "task-1240097" [ 1081.479839] env[62552]: _type = "Task" [ 1081.479839] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.488856] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240097, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.742024] env[62552]: DEBUG nova.scheduler.client.report [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1081.768066] env[62552]: DEBUG oslo_vmware.api [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240093, 'name': PowerOnVM_Task, 'duration_secs': 1.050648} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.768390] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.768599] env[62552]: INFO nova.compute.manager [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Took 7.44 seconds to spawn the instance on the hypervisor. [ 1081.768807] env[62552]: DEBUG nova.compute.manager [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1081.769668] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b4baed-d8f6-4bdc-9371-966fe9ab6854 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.857806] env[62552]: DEBUG oslo_vmware.api [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240096, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197011} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.858275] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.858665] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.859234] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1081.859234] env[62552]: INFO nova.compute.manager [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1081.859605] env[62552]: DEBUG oslo.service.loopingcall [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.859954] env[62552]: DEBUG nova.compute.manager [-] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1081.860094] env[62552]: DEBUG nova.network.neutron [-] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1081.992903] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240097, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.223425] env[62552]: DEBUG nova.compute.manager [req-84d003ef-22b1-4f5f-85f2-d87afdac101c req-aa2d5739-e079-4757-be5c-33b2c8635672 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Received event network-vif-deleted-24b537e1-3264-41fe-8165-63c833fc5c62 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1082.223649] env[62552]: INFO nova.compute.manager [req-84d003ef-22b1-4f5f-85f2-d87afdac101c req-aa2d5739-e079-4757-be5c-33b2c8635672 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Neutron deleted interface 24b537e1-3264-41fe-8165-63c833fc5c62; detaching it from the instance and deleting it from the info cache [ 1082.223846] env[62552]: DEBUG nova.network.neutron [req-84d003ef-22b1-4f5f-85f2-d87afdac101c req-aa2d5739-e079-4757-be5c-33b2c8635672 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.246253] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.248467] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.480s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.248690] env[62552]: DEBUG nova.objects.instance [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'pci_requests' on Instance uuid 158d4227-90ca-41ae-821b-efd353928cb2 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.271295] env[62552]: INFO nova.scheduler.client.report [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Deleted allocations for instance e3f9d5f3-705a-46b1-80d0-bd2421a8b074 [ 1082.292884] env[62552]: INFO nova.compute.manager [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Took 14.92 seconds to build instance. [ 1082.491605] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240097, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.687082] env[62552]: DEBUG nova.network.neutron [-] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.726434] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fbf5b38e-b202-4e99-971e-deff4ea12263 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.736634] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10ff167-c316-4786-b615-c3de1a13b054 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.755127] env[62552]: DEBUG nova.objects.instance [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'numa_topology' on Instance uuid 158d4227-90ca-41ae-821b-efd353928cb2 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.767268] env[62552]: DEBUG nova.compute.manager [req-84d003ef-22b1-4f5f-85f2-d87afdac101c req-aa2d5739-e079-4757-be5c-33b2c8635672 service nova] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Detach interface failed, port_id=24b537e1-3264-41fe-8165-63c833fc5c62, reason: Instance 6c8b5367-4fa0-479b-9382-ff261201e3ef could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1082.779205] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5b8252b0-691c-4159-8091-82a76b22481a tempest-DeleteServersTestJSON-1674271334 tempest-DeleteServersTestJSON-1674271334-project-member] Lock "e3f9d5f3-705a-46b1-80d0-bd2421a8b074" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.432s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.794736] env[62552]: DEBUG oslo_concurrency.lockutils [None req-92112c86-9d1c-480e-8381-cbb750c0d4ae tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.433s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.996375] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240097, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.453085} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.996601] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572/462a3224-d843-45f1-a2bf-69afc18e2572.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1082.996751] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1082.998474] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f5f80bb-0ce8-4320-8a47-3a78cd23c7bb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.009906] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1083.009906] env[62552]: value = "task-1240098" [ 1083.009906] env[62552]: _type = "Task" [ 1083.009906] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.022485] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240098, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.191360] env[62552]: INFO nova.compute.manager [-] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Took 1.33 seconds to deallocate network for instance. [ 1083.262021] env[62552]: INFO nova.compute.claims [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1083.402895] env[62552]: DEBUG nova.compute.manager [req-987c84f4-fe80-4ccd-934b-7ba26dedcafc req-ce07759c-0b5e-4bc7-91f4-fb7e72adf6db service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Received event network-changed-a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1083.402895] env[62552]: DEBUG nova.compute.manager [req-987c84f4-fe80-4ccd-934b-7ba26dedcafc req-ce07759c-0b5e-4bc7-91f4-fb7e72adf6db service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Refreshing instance network info cache due to event network-changed-a1d4afdd-37b2-4b59-b447-0ca443e4c7f6. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1083.404379] env[62552]: DEBUG oslo_concurrency.lockutils [req-987c84f4-fe80-4ccd-934b-7ba26dedcafc req-ce07759c-0b5e-4bc7-91f4-fb7e72adf6db service nova] Acquiring lock "refresh_cache-60238708-4d45-4d73-bd8d-2cbc25704259" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.404789] env[62552]: DEBUG oslo_concurrency.lockutils [req-987c84f4-fe80-4ccd-934b-7ba26dedcafc req-ce07759c-0b5e-4bc7-91f4-fb7e72adf6db service nova] Acquired lock "refresh_cache-60238708-4d45-4d73-bd8d-2cbc25704259" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.405539] env[62552]: DEBUG nova.network.neutron [req-987c84f4-fe80-4ccd-934b-7ba26dedcafc req-ce07759c-0b5e-4bc7-91f4-fb7e72adf6db service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Refreshing network info cache for port a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1083.521876] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240098, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07286} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.522282] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1083.523138] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7677bb4-7885-4791-9c82-893750f20e01 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.550081] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572/462a3224-d843-45f1-a2bf-69afc18e2572.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.551228] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7dcc6296-6c7b-4748-b7c1-9b4f2d290c36 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.576440] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1083.576440] env[62552]: value = "task-1240100" [ 1083.576440] env[62552]: _type = "Task" [ 1083.576440] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.585648] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240100, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.697884] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.087986] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240100, 'name': ReconfigVM_Task, 'duration_secs': 0.496365} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.088378] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572/462a3224-d843-45f1-a2bf-69afc18e2572.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.089473] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'guest_format': None, 'encryption_options': None, 'device_type': 'disk', 'size': 0, 'disk_bus': None, 'boot_index': 0, 'encryption_secret_uuid': None, 'encrypted': False, 'encryption_format': None, 'image_id': '562b48de-e081-4c62-a3e2-cf4ccfd0d1ee'}], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'mount_device': '/dev/sdb', 'delete_on_termination': False, 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'}, 'disk_bus': None, 'boot_index': None, 'attachment_id': '20e0818e-3333-422d-95f3-706dd8f89947', 'volume_type': None}], 'swap': None} {{(pid=62552) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1084.089681] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1084.089874] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1084.090695] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd5331d-065f-41ce-a767-34913d355384 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.108814] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2caf387-6afb-4031-972b-67b55be118a1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.136474] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-90b18b24-f92b-4732-82b5-d40f557a142a/volume-90b18b24-f92b-4732-82b5-d40f557a142a.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.136809] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82f40161-6695-4e0e-b02a-26ac781cf1fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.163135] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1084.163135] env[62552]: value = "task-1240101" [ 1084.163135] env[62552]: _type = "Task" [ 1084.163135] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.171924] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240101, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.408879] env[62552]: DEBUG nova.network.neutron [req-987c84f4-fe80-4ccd-934b-7ba26dedcafc req-ce07759c-0b5e-4bc7-91f4-fb7e72adf6db service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Updated VIF entry in instance network info cache for port a1d4afdd-37b2-4b59-b447-0ca443e4c7f6. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1084.409351] env[62552]: DEBUG nova.network.neutron [req-987c84f4-fe80-4ccd-934b-7ba26dedcafc req-ce07759c-0b5e-4bc7-91f4-fb7e72adf6db service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Updating instance_info_cache with network_info: [{"id": "a1d4afdd-37b2-4b59-b447-0ca443e4c7f6", "address": "fa:16:3e:9f:ce:e5", "network": {"id": "1e4dbebe-8cfc-4a51-8d96-c62d010117eb", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1455446504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3758acde6e3e4320bf63d4cd1667bfc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1d4afdd-37", "ovs_interfaceid": "a1d4afdd-37b2-4b59-b447-0ca443e4c7f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.444829] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2972d2-2355-4bdd-b992-397af68371e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.453708] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b59c2eb-cbab-4eaa-a695-d366c13563cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.487821] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d55e6b-63ee-4cce-b426-b0d7d272b85d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.496451] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a489bdec-02b1-490b-b683-97cec9dd58eb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.511945] env[62552]: DEBUG nova.compute.provider_tree [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.675497] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240101, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.918135] env[62552]: DEBUG oslo_concurrency.lockutils [req-987c84f4-fe80-4ccd-934b-7ba26dedcafc req-ce07759c-0b5e-4bc7-91f4-fb7e72adf6db service nova] Releasing lock "refresh_cache-60238708-4d45-4d73-bd8d-2cbc25704259" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.018979] env[62552]: DEBUG nova.scheduler.client.report [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1085.178780] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240101, 'name': ReconfigVM_Task, 'duration_secs': 0.833126} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.179446] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-90b18b24-f92b-4732-82b5-d40f557a142a/volume-90b18b24-f92b-4732-82b5-d40f557a142a.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1085.184153] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ceb52913-bd82-432e-8035-90614ada577a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.203891] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1085.203891] env[62552]: value = "task-1240102" [ 1085.203891] env[62552]: _type = "Task" [ 1085.203891] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.217374] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240102, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.524832] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.276s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.531541] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.829s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.531541] env[62552]: DEBUG nova.objects.instance [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lazy-loading 'resources' on Instance uuid 6c8b5367-4fa0-479b-9382-ff261201e3ef {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.566300] env[62552]: INFO nova.network.neutron [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updating port dd1028f0-d5a9-48be-84e3-9401cde0a1ff with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1085.716415] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240102, 'name': ReconfigVM_Task, 'duration_secs': 0.370313} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.716934] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1085.717723] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4aa78931-fd42-4bd0-ad65-6a9ca2bc11cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.726998] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1085.726998] env[62552]: value = "task-1240103" [ 1085.726998] env[62552]: _type = "Task" [ 1085.726998] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.740134] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240103, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.102629] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "ed563ce3-40c5-4be0-8948-4d138e81937a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.102948] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "ed563ce3-40c5-4be0-8948-4d138e81937a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.170884] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b4a4fa-06f5-4edb-a03f-3b8974c3811e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.180376] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aeacd77-7679-41a2-b0db-b1760d98af84 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.211683] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a6f0b8-306c-4c56-94c7-ca052c60c20d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.222117] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92518b9-6d4d-44de-874c-cc22908dbd61 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.235160] env[62552]: DEBUG nova.compute.provider_tree [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.244910] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240103, 'name': Rename_Task, 'duration_secs': 0.320569} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.245180] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1086.245475] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a837bb7-f702-40db-924f-662c18a0f53a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.252921] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1086.252921] env[62552]: value = "task-1240104" [ 1086.252921] env[62552]: _type = "Task" [ 1086.252921] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.261174] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240104, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.606156] env[62552]: DEBUG nova.compute.manager [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1086.740077] env[62552]: DEBUG nova.scheduler.client.report [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1086.764290] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240104, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.137469] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.244987] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.718s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.248746] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.111s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.250184] env[62552]: INFO nova.compute.claims [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.274767] env[62552]: DEBUG oslo_vmware.api [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240104, 'name': PowerOnVM_Task, 'duration_secs': 0.872172} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.274767] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1087.274767] env[62552]: DEBUG nova.compute.manager [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1087.275273] env[62552]: INFO nova.scheduler.client.report [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Deleted allocations for instance 6c8b5367-4fa0-479b-9382-ff261201e3ef [ 1087.277582] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac495dd5-57fd-42a2-b614-eb9c455206e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.328833] env[62552]: DEBUG nova.compute.manager [req-abb3acd3-32c4-4e80-9147-d387f468668a req-36d3101a-c823-4064-ac49-4e5089e25562 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received event network-vif-plugged-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1087.329092] env[62552]: DEBUG oslo_concurrency.lockutils [req-abb3acd3-32c4-4e80-9147-d387f468668a req-36d3101a-c823-4064-ac49-4e5089e25562 service nova] Acquiring lock "158d4227-90ca-41ae-821b-efd353928cb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.329358] env[62552]: DEBUG oslo_concurrency.lockutils [req-abb3acd3-32c4-4e80-9147-d387f468668a req-36d3101a-c823-4064-ac49-4e5089e25562 service nova] Lock "158d4227-90ca-41ae-821b-efd353928cb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.329451] env[62552]: DEBUG oslo_concurrency.lockutils [req-abb3acd3-32c4-4e80-9147-d387f468668a req-36d3101a-c823-4064-ac49-4e5089e25562 service nova] Lock "158d4227-90ca-41ae-821b-efd353928cb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.329634] env[62552]: DEBUG nova.compute.manager [req-abb3acd3-32c4-4e80-9147-d387f468668a req-36d3101a-c823-4064-ac49-4e5089e25562 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] No waiting events found dispatching network-vif-plugged-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1087.329782] env[62552]: WARNING nova.compute.manager [req-abb3acd3-32c4-4e80-9147-d387f468668a req-36d3101a-c823-4064-ac49-4e5089e25562 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received unexpected event network-vif-plugged-dd1028f0-d5a9-48be-84e3-9401cde0a1ff for instance with vm_state shelved_offloaded and task_state spawning. [ 1087.487688] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.487688] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.487688] env[62552]: DEBUG nova.network.neutron [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.792204] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dae6625a-e2a2-49e8-af45-846de9f6ed57 tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "6c8b5367-4fa0-479b-9382-ff261201e3ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.575s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.810931] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.236221] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "11831580-1b58-476a-91ce-a4e55947fd91" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.236501] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "11831580-1b58-476a-91ce-a4e55947fd91" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.236720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "11831580-1b58-476a-91ce-a4e55947fd91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.236912] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "11831580-1b58-476a-91ce-a4e55947fd91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.237144] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "11831580-1b58-476a-91ce-a4e55947fd91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.239864] env[62552]: INFO nova.compute.manager [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Terminating instance [ 1088.364322] env[62552]: DEBUG nova.network.neutron [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updating instance_info_cache with network_info: [{"id": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "address": "fa:16:3e:38:6f:9c", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1028f0-d5", "ovs_interfaceid": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.367863] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40a06f6-71c1-43ac-86e9-71f0be2d3f09 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.376019] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b2c29a-0b3e-4e19-a48e-0f5e99afd102 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.408893] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3f23b2-25dd-4e25-a8a9-a6977e93d6d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.416952] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49dfeb01-6e14-4452-98a2-40a2eba71e0d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.431549] env[62552]: DEBUG nova.compute.provider_tree [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.743841] env[62552]: DEBUG nova.compute.manager [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1088.744190] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1088.745112] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1982437e-154a-43ed-b722-8b08f908ba2a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.754084] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1088.754365] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0368018d-b72f-48a4-9040-160210f4008a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.761465] env[62552]: DEBUG oslo_vmware.api [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1088.761465] env[62552]: value = "task-1240105" [ 1088.761465] env[62552]: _type = "Task" [ 1088.761465] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.770322] env[62552]: DEBUG oslo_vmware.api [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.866919] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.897978] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5a9a21726434818a2db838f7a071cd10',container_format='bare',created_at=2024-10-10T11:39:10Z,direct_url=,disk_format='vmdk',id=6b29c2df-72f4-4913-904a-2589eab9d481,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-831025572-shelved',owner='25030025e90243b290a7d90efa26ec79',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-10T11:39:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1088.898410] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1088.898649] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1088.899263] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1088.899263] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1088.899374] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1088.899584] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1088.899759] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1088.899930] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1088.900161] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1088.900314] env[62552]: DEBUG nova.virt.hardware [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1088.901236] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531ff956-41e8-4d42-879c-ea2569a76a0a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.911336] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad286d09-945e-4fae-accf-f661234c77e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.930174] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:6f:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd1028f0-d5a9-48be-84e3-9401cde0a1ff', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1088.939148] env[62552]: DEBUG oslo.service.loopingcall [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.940380] env[62552]: DEBUG nova.scheduler.client.report [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1088.944266] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1088.944908] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b1fd2ac-2fcf-4b85-8cd2-647dd374b651 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.967640] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1088.967640] env[62552]: value = "task-1240106" [ 1088.967640] env[62552]: _type = "Task" [ 1088.967640] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.977081] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240106, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.273040] env[62552]: DEBUG oslo_vmware.api [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240105, 'name': PowerOffVM_Task, 'duration_secs': 0.221483} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.273040] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1089.273312] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1089.273433] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a75bf73-83fc-44bc-857d-63fc9b4be753 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.344253] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1089.344633] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1089.344903] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Deleting the datastore file [datastore1] 11831580-1b58-476a-91ce-a4e55947fd91 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.345309] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be858e38-16e5-419b-b58a-5641232a32d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.352962] env[62552]: DEBUG oslo_vmware.api [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for the task: (returnval){ [ 1089.352962] env[62552]: value = "task-1240108" [ 1089.352962] env[62552]: _type = "Task" [ 1089.352962] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.357836] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "462a3224-d843-45f1-a2bf-69afc18e2572" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.358194] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.360572] env[62552]: DEBUG nova.compute.manager [req-3c400ffa-c737-457f-9333-7cfddf6f25ba req-c635d04e-03ee-458a-afb4-f9cd20cc39c3 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received event network-changed-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1089.360809] env[62552]: DEBUG nova.compute.manager [req-3c400ffa-c737-457f-9333-7cfddf6f25ba req-c635d04e-03ee-458a-afb4-f9cd20cc39c3 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Refreshing instance network info cache due to event network-changed-dd1028f0-d5a9-48be-84e3-9401cde0a1ff. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1089.361091] env[62552]: DEBUG oslo_concurrency.lockutils [req-3c400ffa-c737-457f-9333-7cfddf6f25ba req-c635d04e-03ee-458a-afb4-f9cd20cc39c3 service nova] Acquiring lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.361320] env[62552]: DEBUG oslo_concurrency.lockutils [req-3c400ffa-c737-457f-9333-7cfddf6f25ba req-c635d04e-03ee-458a-afb4-f9cd20cc39c3 service nova] Acquired lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.361541] env[62552]: DEBUG nova.network.neutron [req-3c400ffa-c737-457f-9333-7cfddf6f25ba req-c635d04e-03ee-458a-afb4-f9cd20cc39c3 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Refreshing network info cache for port dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1089.369382] env[62552]: DEBUG oslo_vmware.api [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240108, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.446192] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.446984] env[62552]: DEBUG nova.compute.manager [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1089.450778] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.640s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.451084] env[62552]: DEBUG nova.objects.instance [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62552) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1089.477682] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240106, 'name': CreateVM_Task, 'duration_secs': 0.427411} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.477893] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1089.478562] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.478744] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.479141] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1089.479415] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88858f36-314f-4af8-8f41-eaa66c2fb8e0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.484603] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1089.484603] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527f732c-717a-f9af-8db9-465e79df7ec1" [ 1089.484603] env[62552]: _type = "Task" [ 1089.484603] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.494496] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]527f732c-717a-f9af-8db9-465e79df7ec1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.865592] env[62552]: INFO nova.compute.manager [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Detaching volume 90b18b24-f92b-4732-82b5-d40f557a142a [ 1089.869597] env[62552]: DEBUG oslo_vmware.api [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Task: {'id': task-1240108, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.295456} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.871026] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1089.871026] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1089.871026] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1089.871211] env[62552]: INFO nova.compute.manager [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1089.871775] env[62552]: DEBUG oslo.service.loopingcall [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.872390] env[62552]: DEBUG nova.compute.manager [-] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1089.872490] env[62552]: DEBUG nova.network.neutron [-] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1089.903143] env[62552]: INFO nova.virt.block_device [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Attempting to driver detach volume 90b18b24-f92b-4732-82b5-d40f557a142a from mountpoint /dev/sdb [ 1089.903455] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1089.903741] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1089.904735] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0859e42d-4a33-401b-b407-fd37f4d9e29c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.934105] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3947fc1-aa85-461b-86ad-007403493b28 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.942097] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f849f036-83e4-40d8-94aa-5cac6bd6f509 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.964138] env[62552]: DEBUG nova.compute.utils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1089.968367] env[62552]: DEBUG nova.compute.manager [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1089.968571] env[62552]: DEBUG nova.network.neutron [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1089.970790] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a6fc21-81e3-4698-8f6e-60bc2de59aa7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.987982] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] The volume has not been displaced from its original location: [datastore1] volume-90b18b24-f92b-4732-82b5-d40f557a142a/volume-90b18b24-f92b-4732-82b5-d40f557a142a.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1089.993269] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1089.997418] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af758552-5faa-41f5-91d2-64c7393f676c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.013011] env[62552]: DEBUG nova.policy [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e28ac6cf2a540b7afd48e77f8d299c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1223fc0363644d1b964ca7f581e9c771', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1090.022960] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.022960] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Processing image 6b29c2df-72f4-4913-904a-2589eab9d481 {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1090.022960] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481/6b29c2df-72f4-4913-904a-2589eab9d481.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.022960] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481/6b29c2df-72f4-4913-904a-2589eab9d481.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.022960] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1090.023216] env[62552]: DEBUG oslo_vmware.api [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1090.023216] env[62552]: value = "task-1240109" [ 1090.023216] env[62552]: _type = "Task" [ 1090.023216] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.023331] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b943d76e-800f-418d-9b21-9505f0c8a20e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.033923] env[62552]: DEBUG oslo_vmware.api [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240109, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.036683] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1090.036855] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1090.037591] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f100e021-c3ae-4944-a708-59e299d21779 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.043212] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1090.043212] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525d0736-57d7-6609-ce38-fdb972216829" [ 1090.043212] env[62552]: _type = "Task" [ 1090.043212] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.052296] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525d0736-57d7-6609-ce38-fdb972216829, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.304686] env[62552]: DEBUG nova.network.neutron [req-3c400ffa-c737-457f-9333-7cfddf6f25ba req-c635d04e-03ee-458a-afb4-f9cd20cc39c3 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updated VIF entry in instance network info cache for port dd1028f0-d5a9-48be-84e3-9401cde0a1ff. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1090.305958] env[62552]: DEBUG nova.network.neutron [req-3c400ffa-c737-457f-9333-7cfddf6f25ba req-c635d04e-03ee-458a-afb4-f9cd20cc39c3 service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updating instance_info_cache with network_info: [{"id": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "address": "fa:16:3e:38:6f:9c", "network": {"id": "857e1f0c-1ed4-4816-84e7-dc5b12b42518", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1259168931-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25030025e90243b290a7d90efa26ec79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1028f0-d5", "ovs_interfaceid": "dd1028f0-d5a9-48be-84e3-9401cde0a1ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.365114] env[62552]: DEBUG nova.network.neutron [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Successfully created port: 00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1090.469684] env[62552]: DEBUG nova.compute.manager [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1090.475570] env[62552]: DEBUG oslo_concurrency.lockutils [None req-2bb0d7ec-0e1f-4f25-b9b2-eb11cd711d23 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.025s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.535519] env[62552]: DEBUG oslo_vmware.api [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240109, 'name': ReconfigVM_Task, 'duration_secs': 0.404196} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.535813] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1090.540315] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91d10035-e98e-4a4d-a1fb-b151ec82fb0a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.560073] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Preparing fetch location {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1090.560344] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Fetch image to [datastore2] OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c/OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c.vmdk {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1090.560531] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Downloading stream optimized image 6b29c2df-72f4-4913-904a-2589eab9d481 to [datastore2] OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c/OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c.vmdk on the data store datastore2 as vApp {{(pid=62552) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1090.560702] env[62552]: DEBUG nova.virt.vmwareapi.images [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Downloading image file data 6b29c2df-72f4-4913-904a-2589eab9d481 to the ESX as VM named 'OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c' {{(pid=62552) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1090.562626] env[62552]: DEBUG oslo_vmware.api [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1090.562626] env[62552]: value = "task-1240110" [ 1090.562626] env[62552]: _type = "Task" [ 1090.562626] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.574063] env[62552]: DEBUG oslo_vmware.api [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240110, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.640414] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1090.640414] env[62552]: value = "resgroup-9" [ 1090.640414] env[62552]: _type = "ResourcePool" [ 1090.640414] env[62552]: }. {{(pid=62552) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1090.640777] env[62552]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-67eb6040-06c7-417e-ba9f-4b03ce1cc736 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.664691] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lease: (returnval){ [ 1090.664691] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523e02b0-5a62-7ac3-c512-5eabe4846e13" [ 1090.664691] env[62552]: _type = "HttpNfcLease" [ 1090.664691] env[62552]: } obtained for vApp import into resource pool (val){ [ 1090.664691] env[62552]: value = "resgroup-9" [ 1090.664691] env[62552]: _type = "ResourcePool" [ 1090.664691] env[62552]: }. {{(pid=62552) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1090.665279] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the lease: (returnval){ [ 1090.665279] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523e02b0-5a62-7ac3-c512-5eabe4846e13" [ 1090.665279] env[62552]: _type = "HttpNfcLease" [ 1090.665279] env[62552]: } to be ready. {{(pid=62552) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1090.672033] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1090.672033] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523e02b0-5a62-7ac3-c512-5eabe4846e13" [ 1090.672033] env[62552]: _type = "HttpNfcLease" [ 1090.672033] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1090.804340] env[62552]: DEBUG nova.network.neutron [-] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.808706] env[62552]: DEBUG oslo_concurrency.lockutils [req-3c400ffa-c737-457f-9333-7cfddf6f25ba req-c635d04e-03ee-458a-afb4-f9cd20cc39c3 service nova] Releasing lock "refresh_cache-158d4227-90ca-41ae-821b-efd353928cb2" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.073884] env[62552]: DEBUG oslo_vmware.api [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240110, 'name': ReconfigVM_Task, 'duration_secs': 0.222134} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.074419] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267522', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'name': 'volume-90b18b24-f92b-4732-82b5-d40f557a142a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '462a3224-d843-45f1-a2bf-69afc18e2572', 'attached_at': '', 'detached_at': '', 'volume_id': '90b18b24-f92b-4732-82b5-d40f557a142a', 'serial': '90b18b24-f92b-4732-82b5-d40f557a142a'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1091.175320] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1091.175320] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523e02b0-5a62-7ac3-c512-5eabe4846e13" [ 1091.175320] env[62552]: _type = "HttpNfcLease" [ 1091.175320] env[62552]: } is initializing. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1091.308053] env[62552]: INFO nova.compute.manager [-] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Took 1.43 seconds to deallocate network for instance. [ 1091.382415] env[62552]: DEBUG nova.compute.manager [req-979b6b1c-3fd4-4ad4-8b4d-99e0903be473 req-8e26d3c5-c39b-4eff-8b8f-1ca7eef7a97b service nova] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Received event network-vif-deleted-123341c8-40fb-4f81-9d1c-1f556b6a92b8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1091.480451] env[62552]: DEBUG nova.compute.manager [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1091.485192] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.485452] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.504167] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.504454] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.504621] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.504809] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.504959] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.505127] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.505340] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.505506] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.505685] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.506019] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.506019] env[62552]: DEBUG nova.virt.hardware [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.507292] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95417327-f6db-457b-8a66-0d0d58aaffbb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.516316] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8411743-aeca-4b18-b69c-4cc24f50b5e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.618659] env[62552]: DEBUG nova.objects.instance [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lazy-loading 'flavor' on Instance uuid 462a3224-d843-45f1-a2bf-69afc18e2572 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.674184] env[62552]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1091.674184] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523e02b0-5a62-7ac3-c512-5eabe4846e13" [ 1091.674184] env[62552]: _type = "HttpNfcLease" [ 1091.674184] env[62552]: } is ready. {{(pid=62552) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1091.674558] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1091.674558] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]523e02b0-5a62-7ac3-c512-5eabe4846e13" [ 1091.674558] env[62552]: _type = "HttpNfcLease" [ 1091.674558] env[62552]: }. {{(pid=62552) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1091.675467] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed98ffb-4367-42d5-a7c1-9bd3ad40f220 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.684866] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d17589-8978-a9ad-9c72-a982ce2013bc/disk-0.vmdk from lease info. {{(pid=62552) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1091.685118] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d17589-8978-a9ad-9c72-a982ce2013bc/disk-0.vmdk. {{(pid=62552) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1091.754703] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2e81cfbc-acb8-4783-9ff8-d8237f54a202 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.770492] env[62552]: DEBUG nova.compute.manager [req-e01697dc-9bd0-4cec-ac85-2f9762848cd4 req-59363af9-a00d-4e35-85e6-026190119067 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Received event network-vif-plugged-00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1091.770723] env[62552]: DEBUG oslo_concurrency.lockutils [req-e01697dc-9bd0-4cec-ac85-2f9762848cd4 req-59363af9-a00d-4e35-85e6-026190119067 service nova] Acquiring lock "ed563ce3-40c5-4be0-8948-4d138e81937a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.770941] env[62552]: DEBUG oslo_concurrency.lockutils [req-e01697dc-9bd0-4cec-ac85-2f9762848cd4 req-59363af9-a00d-4e35-85e6-026190119067 service nova] Lock "ed563ce3-40c5-4be0-8948-4d138e81937a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.771105] env[62552]: DEBUG oslo_concurrency.lockutils [req-e01697dc-9bd0-4cec-ac85-2f9762848cd4 req-59363af9-a00d-4e35-85e6-026190119067 service nova] Lock "ed563ce3-40c5-4be0-8948-4d138e81937a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.771277] env[62552]: DEBUG nova.compute.manager [req-e01697dc-9bd0-4cec-ac85-2f9762848cd4 req-59363af9-a00d-4e35-85e6-026190119067 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] No waiting events found dispatching network-vif-plugged-00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1091.771447] env[62552]: WARNING nova.compute.manager [req-e01697dc-9bd0-4cec-ac85-2f9762848cd4 req-59363af9-a00d-4e35-85e6-026190119067 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Received unexpected event network-vif-plugged-00e0a1b2-3fba-44a1-a14d-9846dd2d281d for instance with vm_state building and task_state spawning. [ 1091.814524] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.814776] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.814996] env[62552]: DEBUG nova.objects.instance [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lazy-loading 'resources' on Instance uuid 11831580-1b58-476a-91ce-a4e55947fd91 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.859942] env[62552]: DEBUG nova.network.neutron [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Successfully updated port: 00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1091.988454] env[62552]: DEBUG nova.compute.utils [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1092.361794] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.361945] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquired lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.362121] env[62552]: DEBUG nova.network.neutron [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1092.440510] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97906a24-1c48-4c03-bd25-bbabd38a10ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.451755] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b9bfbe-1489-4afa-b8f4-62dfe57723a2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.489121] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1651aa-8252-4838-8460-3975eb42d9c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.495401] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.501987] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dda79a1-7439-4ea4-a33f-0fc660ce8347 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.519192] env[62552]: DEBUG nova.compute.provider_tree [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.627083] env[62552]: DEBUG oslo_concurrency.lockutils [None req-ae17fae2-69ee-4807-b9d6-e179ee1939fc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.269s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.815959] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Completed reading data from the image iterator. {{(pid=62552) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1092.816269] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d17589-8978-a9ad-9c72-a982ce2013bc/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1092.817207] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23b1d52-1511-4c54-826e-c29eda3eeb23 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.824196] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d17589-8978-a9ad-9c72-a982ce2013bc/disk-0.vmdk is in state: ready. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1092.824365] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d17589-8978-a9ad-9c72-a982ce2013bc/disk-0.vmdk. {{(pid=62552) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1092.824603] env[62552]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ac30439a-d34b-4c19-b2d0-cf0bd13e0759 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.896212] env[62552]: DEBUG nova.network.neutron [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1093.019519] env[62552]: DEBUG nova.network.neutron [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updating instance_info_cache with network_info: [{"id": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "address": "fa:16:3e:42:30:1b", "network": {"id": "4249dd40-74ec-4271-8d53-48d956819e64", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-928492516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1223fc0363644d1b964ca7f581e9c771", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb18870e-f482-4c7b-8cd4-5c933d3ad294", "external-id": "nsx-vlan-transportzone-76", "segmentation_id": 76, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00e0a1b2-3f", "ovs_interfaceid": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.025681] env[62552]: DEBUG nova.scheduler.client.report [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1093.030751] env[62552]: DEBUG oslo_vmware.rw_handles [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d17589-8978-a9ad-9c72-a982ce2013bc/disk-0.vmdk. {{(pid=62552) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1093.031045] env[62552]: INFO nova.virt.vmwareapi.images [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Downloaded image file data 6b29c2df-72f4-4913-904a-2589eab9d481 [ 1093.032108] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7496b6ad-4d30-42b9-9eae-bcfe0b42404b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.049785] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52363293-5025-4359-984f-37907c8e3a52 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.077783] env[62552]: INFO nova.virt.vmwareapi.images [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] The imported VM was unregistered [ 1093.080225] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Caching image {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1093.080476] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Creating directory with path [datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481 {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1093.081073] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32aaa2e1-b6d7-4f5d-b16e-32eba828efec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.093145] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Created directory with path [datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481 {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1093.093374] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c/OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c.vmdk to [datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481/6b29c2df-72f4-4913-904a-2589eab9d481.vmdk. {{(pid=62552) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1093.093644] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-1ab80799-fa33-417d-aab6-985eac9e0ef0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.101878] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1093.101878] env[62552]: value = "task-1240113" [ 1093.101878] env[62552]: _type = "Task" [ 1093.101878] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.110684] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240113, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.523080] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Releasing lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.523080] env[62552]: DEBUG nova.compute.manager [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Instance network_info: |[{"id": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "address": "fa:16:3e:42:30:1b", "network": {"id": "4249dd40-74ec-4271-8d53-48d956819e64", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-928492516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1223fc0363644d1b964ca7f581e9c771", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb18870e-f482-4c7b-8cd4-5c933d3ad294", "external-id": "nsx-vlan-transportzone-76", "segmentation_id": 76, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00e0a1b2-3f", "ovs_interfaceid": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1093.523743] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:30:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb18870e-f482-4c7b-8cd4-5c933d3ad294', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00e0a1b2-3fba-44a1-a14d-9846dd2d281d', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1093.539258] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Creating folder: Project (1223fc0363644d1b964ca7f581e9c771). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1093.540606] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.726s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.544109] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ef92a37d-98a4-4aa8-8c91-df93a9bd863b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.565222] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.565755] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.566217] env[62552]: INFO nova.compute.manager [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Attaching volume 65d3ba2b-bf26-458c-a32c-5eaf30bf1a81 to /dev/sdb [ 1093.568964] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Created folder: Project (1223fc0363644d1b964ca7f581e9c771) in parent group-v267339. [ 1093.569341] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Creating folder: Instances. Parent ref: group-v267527. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1093.571010] env[62552]: INFO nova.scheduler.client.report [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Deleted allocations for instance 11831580-1b58-476a-91ce-a4e55947fd91 [ 1093.574051] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1bf5894-ee2c-4b55-9957-8601b95dd703 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.597677] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Created folder: Instances in parent group-v267527. [ 1093.597949] env[62552]: DEBUG oslo.service.loopingcall [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1093.598194] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1093.598429] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f668b96e-cdfb-4aff-b2b1-e514c0508f2b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.615183] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181099dd-2edf-4f7a-b36d-b2090b7a7ef7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.627614] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240113, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.630248] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1093.630248] env[62552]: value = "task-1240116" [ 1093.630248] env[62552]: _type = "Task" [ 1093.630248] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.631024] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900f2d91-5178-40a5-a225-1ee1060169ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.647012] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240116, 'name': CreateVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.654551] env[62552]: DEBUG nova.virt.block_device [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Updating existing volume attachment record: ab98b8b7-e5c5-46e9-a118-f72f940c2e04 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1093.703061] env[62552]: DEBUG oslo_concurrency.lockutils [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "462a3224-d843-45f1-a2bf-69afc18e2572" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.703061] env[62552]: DEBUG oslo_concurrency.lockutils [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.703347] env[62552]: DEBUG oslo_concurrency.lockutils [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "462a3224-d843-45f1-a2bf-69afc18e2572-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.703590] env[62552]: DEBUG oslo_concurrency.lockutils [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.703971] env[62552]: DEBUG oslo_concurrency.lockutils [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.707057] env[62552]: INFO nova.compute.manager [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Terminating instance [ 1093.798466] env[62552]: DEBUG nova.compute.manager [req-9a839607-d3c1-478c-b887-02b1e10d27f5 req-de53caea-5de8-4e86-89fb-935e02ec30f6 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Received event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1093.798706] env[62552]: DEBUG nova.compute.manager [req-9a839607-d3c1-478c-b887-02b1e10d27f5 req-de53caea-5de8-4e86-89fb-935e02ec30f6 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing instance network info cache due to event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1093.799106] env[62552]: DEBUG oslo_concurrency.lockutils [req-9a839607-d3c1-478c-b887-02b1e10d27f5 req-de53caea-5de8-4e86-89fb-935e02ec30f6 service nova] Acquiring lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.799191] env[62552]: DEBUG oslo_concurrency.lockutils [req-9a839607-d3c1-478c-b887-02b1e10d27f5 req-de53caea-5de8-4e86-89fb-935e02ec30f6 service nova] Acquired lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.799417] env[62552]: DEBUG nova.network.neutron [req-9a839607-d3c1-478c-b887-02b1e10d27f5 req-de53caea-5de8-4e86-89fb-935e02ec30f6 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1094.088322] env[62552]: DEBUG oslo_concurrency.lockutils [None req-fadac258-08a9-4cd2-87ed-bac5100e6dca tempest-ServerRescueNegativeTestJSON-880604856 tempest-ServerRescueNegativeTestJSON-880604856-project-member] Lock "11831580-1b58-476a-91ce-a4e55947fd91" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.852s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.128523] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240113, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.146816] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240116, 'name': CreateVM_Task, 'duration_secs': 0.453415} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.147025] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1094.147829] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.148093] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.148488] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1094.148806] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b716436-325c-4318-ba6f-062ffa6c6ca1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.156734] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1094.156734] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52729d32-92c3-254d-44c6-9f4b7d20ea7d" [ 1094.156734] env[62552]: _type = "Task" [ 1094.156734] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.168736] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52729d32-92c3-254d-44c6-9f4b7d20ea7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.211900] env[62552]: DEBUG nova.compute.manager [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1094.212156] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1094.213264] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1542d30c-fd06-4343-b25d-20e6a42ecbe7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.224249] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1094.224556] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9815cca-664e-43c2-b5aa-6b2289989377 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.235756] env[62552]: DEBUG oslo_vmware.api [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1094.235756] env[62552]: value = "task-1240118" [ 1094.235756] env[62552]: _type = "Task" [ 1094.235756] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.247646] env[62552]: DEBUG oslo_vmware.api [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240118, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.520143] env[62552]: DEBUG nova.network.neutron [req-9a839607-d3c1-478c-b887-02b1e10d27f5 req-de53caea-5de8-4e86-89fb-935e02ec30f6 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updated VIF entry in instance network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1094.520596] env[62552]: DEBUG nova.network.neutron [req-9a839607-d3c1-478c-b887-02b1e10d27f5 req-de53caea-5de8-4e86-89fb-935e02ec30f6 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updating instance_info_cache with network_info: [{"id": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "address": "fa:16:3e:42:30:1b", "network": {"id": "4249dd40-74ec-4271-8d53-48d956819e64", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-928492516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1223fc0363644d1b964ca7f581e9c771", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb18870e-f482-4c7b-8cd4-5c933d3ad294", "external-id": "nsx-vlan-transportzone-76", "segmentation_id": 76, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00e0a1b2-3f", "ovs_interfaceid": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.627103] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240113, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.669735] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52729d32-92c3-254d-44c6-9f4b7d20ea7d, 'name': SearchDatastore_Task, 'duration_secs': 0.013942} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.670018] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.670310] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1094.670579] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.670787] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.671087] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1094.671445] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24ee8dd7-0080-4595-81a2-2c05d75a12cd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.685037] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1094.685403] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1094.686266] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34d58b06-dcd1-4f10-8e03-5d4500758c10 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.693793] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1094.693793] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52060526-ee22-82a5-e336-f788b41f6028" [ 1094.693793] env[62552]: _type = "Task" [ 1094.693793] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.698284] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.698421] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Cleaning up deleted instances {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11571}} [ 1094.704957] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52060526-ee22-82a5-e336-f788b41f6028, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.748460] env[62552]: DEBUG oslo_vmware.api [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240118, 'name': PowerOffVM_Task, 'duration_secs': 0.500901} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.748882] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1094.749082] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1094.749384] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7f7dcd3f-a9dd-4785-83bf-438b3c644539 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.828042] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1094.828042] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1094.828042] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleting the datastore file [datastore2] 462a3224-d843-45f1-a2bf-69afc18e2572 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1094.828042] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35110415-668d-40fa-9ca9-babbb66351d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.837680] env[62552]: DEBUG oslo_vmware.api [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1094.837680] env[62552]: value = "task-1240120" [ 1094.837680] env[62552]: _type = "Task" [ 1094.837680] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.848239] env[62552]: DEBUG oslo_vmware.api [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240120, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.024563] env[62552]: DEBUG oslo_concurrency.lockutils [req-9a839607-d3c1-478c-b887-02b1e10d27f5 req-de53caea-5de8-4e86-89fb-935e02ec30f6 service nova] Releasing lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.129450] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240113, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.212835] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] There are 53 instances to clean {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11580}} [ 1095.213073] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6824a837-60d0-4577-927f-a54d28a68f80] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1095.219381] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52060526-ee22-82a5-e336-f788b41f6028, 'name': SearchDatastore_Task, 'duration_secs': 0.013581} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.220464] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec5fb774-af5e-4020-9cee-5ba3ba3b9b63 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.229918] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1095.229918] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c348ab-bcef-5a4e-a8d1-c6a218f4358d" [ 1095.229918] env[62552]: _type = "Task" [ 1095.229918] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.240716] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c348ab-bcef-5a4e-a8d1-c6a218f4358d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.352117] env[62552]: DEBUG oslo_vmware.api [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240120, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.627959] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240113, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.371325} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.628277] env[62552]: INFO nova.virt.vmwareapi.ds_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c/OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c.vmdk to [datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481/6b29c2df-72f4-4913-904a-2589eab9d481.vmdk. [ 1095.628452] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Cleaning up location [datastore2] OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1095.628615] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_097f6469-ee08-4ed5-9aaf-4211d634d47c {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1095.628875] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9bed553a-6dce-4226-a10d-5cfaf01718a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.637023] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1095.637023] env[62552]: value = "task-1240121" [ 1095.637023] env[62552]: _type = "Task" [ 1095.637023] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.645236] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240121, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.716218] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 7e319ad7-ca42-4eb1-b303-88adbfba7bfe] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1095.740819] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c348ab-bcef-5a4e-a8d1-c6a218f4358d, 'name': SearchDatastore_Task, 'duration_secs': 0.013563} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.741042] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.741303] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] ed563ce3-40c5-4be0-8948-4d138e81937a/ed563ce3-40c5-4be0-8948-4d138e81937a.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1095.741683] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e70f1300-63a3-4984-95a7-3fe63aed78b4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.750399] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1095.750399] env[62552]: value = "task-1240122" [ 1095.750399] env[62552]: _type = "Task" [ 1095.750399] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.761653] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240122, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.848713] env[62552]: DEBUG oslo_vmware.api [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240120, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.679971} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.848974] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1095.849188] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1095.849371] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1095.849548] env[62552]: INFO nova.compute.manager [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1095.849785] env[62552]: DEBUG oslo.service.loopingcall [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1095.849980] env[62552]: DEBUG nova.compute.manager [-] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1095.850091] env[62552]: DEBUG nova.network.neutron [-] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1096.149334] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240121, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.063185} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.150212] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.150580] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481/6b29c2df-72f4-4913-904a-2589eab9d481.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.150988] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481/6b29c2df-72f4-4913-904a-2589eab9d481.vmdk to [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2/158d4227-90ca-41ae-821b-efd353928cb2.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1096.152278] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f13cebb6-559b-4f9e-b2d5-9823e946b637 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.163023] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1096.163023] env[62552]: value = "task-1240124" [ 1096.163023] env[62552]: _type = "Task" [ 1096.163023] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.179604] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.219523] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: e3f9d5f3-705a-46b1-80d0-bd2421a8b074] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1096.262411] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240122, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494569} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.263128] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] ed563ce3-40c5-4be0-8948-4d138e81937a/ed563ce3-40c5-4be0-8948-4d138e81937a.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1096.263376] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1096.263642] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7a789a5-415a-48a2-9183-a5934b6c57bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.275405] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1096.275405] env[62552]: value = "task-1240125" [ 1096.275405] env[62552]: _type = "Task" [ 1096.275405] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.285140] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240125, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.526372] env[62552]: DEBUG nova.compute.manager [req-ad63699b-4cff-42a0-80bd-4bebf748dfa5 req-62d5fa4d-9009-45a1-b1f1-bd3fb48dad8b service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Received event network-vif-deleted-81435491-9993-42c8-80d8-8c4f23693ce4 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1096.526474] env[62552]: INFO nova.compute.manager [req-ad63699b-4cff-42a0-80bd-4bebf748dfa5 req-62d5fa4d-9009-45a1-b1f1-bd3fb48dad8b service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Neutron deleted interface 81435491-9993-42c8-80d8-8c4f23693ce4; detaching it from the instance and deleting it from the info cache [ 1096.526647] env[62552]: DEBUG nova.network.neutron [req-ad63699b-4cff-42a0-80bd-4bebf748dfa5 req-62d5fa4d-9009-45a1-b1f1-bd3fb48dad8b service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.673398] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240124, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.725694] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 19b69199-b3e1-45c5-af35-d9329113d360] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1096.789133] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240125, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075356} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.790212] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1096.791814] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6824b74f-fb5c-429a-9f4e-a7082f3328aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.819539] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] ed563ce3-40c5-4be0-8948-4d138e81937a/ed563ce3-40c5-4be0-8948-4d138e81937a.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.820912] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95bf2ba4-6334-4723-a5bc-6656527c8974 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.844414] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1096.844414] env[62552]: value = "task-1240126" [ 1096.844414] env[62552]: _type = "Task" [ 1096.844414] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.856846] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240126, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.000544] env[62552]: DEBUG nova.network.neutron [-] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.030039] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6acb0fa5-f2e8-4a88-8599-18c92847cfec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.044326] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a63ac2a-6102-4ba4-8dcf-bbf0b351cccb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.079430] env[62552]: DEBUG nova.compute.manager [req-ad63699b-4cff-42a0-80bd-4bebf748dfa5 req-62d5fa4d-9009-45a1-b1f1-bd3fb48dad8b service nova] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Detach interface failed, port_id=81435491-9993-42c8-80d8-8c4f23693ce4, reason: Instance 462a3224-d843-45f1-a2bf-69afc18e2572 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1097.175513] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240124, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.229410] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 2f30cf85-df40-4560-8cc3-d0cdf7cbecc5] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1097.357041] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240126, 'name': ReconfigVM_Task, 'duration_secs': 0.339425} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.357296] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Reconfigured VM instance instance-0000006a to attach disk [datastore1] ed563ce3-40c5-4be0-8948-4d138e81937a/ed563ce3-40c5-4be0-8948-4d138e81937a.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1097.357981] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef9a3e3b-251b-48ce-8519-3e71c0ba0297 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.367593] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1097.367593] env[62552]: value = "task-1240127" [ 1097.367593] env[62552]: _type = "Task" [ 1097.367593] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.381010] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240127, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.503433] env[62552]: INFO nova.compute.manager [-] [instance: 462a3224-d843-45f1-a2bf-69afc18e2572] Took 1.65 seconds to deallocate network for instance. [ 1097.676676] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240124, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.733047] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 16785811-1927-408b-9ead-5848a5ffdf24] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1097.880836] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240127, 'name': Rename_Task, 'duration_secs': 0.186425} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.881115] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1097.881401] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-075e317e-d087-4a7a-912f-b7d982b76331 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.890883] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1097.890883] env[62552]: value = "task-1240128" [ 1097.890883] env[62552]: _type = "Task" [ 1097.890883] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.900898] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240128, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.011567] env[62552]: DEBUG oslo_concurrency.lockutils [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.011907] env[62552]: DEBUG oslo_concurrency.lockutils [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.012229] env[62552]: DEBUG nova.objects.instance [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lazy-loading 'resources' on Instance uuid 462a3224-d843-45f1-a2bf-69afc18e2572 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.176541] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240124, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.204514] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1098.204758] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267530', 'volume_id': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'name': 'volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '765bbf00-2569-4c49-9634-81d10f375175', 'attached_at': '', 'detached_at': '', 'volume_id': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'serial': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1098.205694] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02007e0c-c4ca-48df-801e-afbc041640ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.224928] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d932dc-af4a-44e4-9a9a-9e4616a6228d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.245811] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 38bf968c-e757-4b46-b069-f625f5041c84] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1098.256068] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81/volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.256396] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8601487c-b083-48a6-8ab1-f06fef16c678 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.280142] env[62552]: DEBUG oslo_vmware.api [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1098.280142] env[62552]: value = "task-1240129" [ 1098.280142] env[62552]: _type = "Task" [ 1098.280142] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.294132] env[62552]: DEBUG oslo_vmware.api [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240129, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.404213] env[62552]: DEBUG oslo_vmware.api [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240128, 'name': PowerOnVM_Task, 'duration_secs': 0.498596} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.404526] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1098.404742] env[62552]: INFO nova.compute.manager [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Took 6.92 seconds to spawn the instance on the hypervisor. [ 1098.404933] env[62552]: DEBUG nova.compute.manager [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1098.405817] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4863ca4-8ee5-4c63-a8e5-59852e2ec160 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.623865] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570c80ce-f300-45cc-951f-782af0e73a37 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.632405] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45543c7-f92d-4392-8716-e3a1f5f685cb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.663485] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50ae140-b444-4807-8206-add0e5a01f8e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.676369] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b452024e-cb2f-4e71-8513-1236b9232a71 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.680016] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240124, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.483231} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.680359] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6b29c2df-72f4-4913-904a-2589eab9d481/6b29c2df-72f4-4913-904a-2589eab9d481.vmdk to [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2/158d4227-90ca-41ae-821b-efd353928cb2.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1098.681364] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3639c916-a6b4-46d9-ad99-97135d322d7b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.691196] env[62552]: DEBUG nova.compute.provider_tree [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.713146] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2/158d4227-90ca-41ae-821b-efd353928cb2.vmdk or device None with type streamOptimized {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.714095] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b09ca44-027a-4a3e-8d35-77eb5fb26941 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.735392] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1098.735392] env[62552]: value = "task-1240130" [ 1098.735392] env[62552]: _type = "Task" [ 1098.735392] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.747316] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240130, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.748865] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 0ed1e5cc-d5c7-4d24-8a79-6e498b1229c9] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1098.792432] env[62552]: DEBUG oslo_vmware.api [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240129, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.929810] env[62552]: INFO nova.compute.manager [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Took 11.82 seconds to build instance. [ 1099.194928] env[62552]: DEBUG nova.scheduler.client.report [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1099.245683] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240130, 'name': ReconfigVM_Task, 'duration_secs': 0.302212} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.246107] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2/158d4227-90ca-41ae-821b-efd353928cb2.vmdk or device None with type streamOptimized {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.246796] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78f0b543-9e97-408c-b5e0-054955b4714d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.251986] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: c47b92b5-a6e5-4ea4-bdc9-beba16c3ce0e] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1099.255041] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1099.255041] env[62552]: value = "task-1240131" [ 1099.255041] env[62552]: _type = "Task" [ 1099.255041] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.263157] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240131, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.292401] env[62552]: DEBUG oslo_vmware.api [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240129, 'name': ReconfigVM_Task, 'duration_secs': 0.58329} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.292801] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfigured VM instance instance-00000065 to attach disk [datastore1] volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81/volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.298053] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54120f7d-59da-4340-86ce-75906e3888d7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.315055] env[62552]: DEBUG oslo_vmware.api [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1099.315055] env[62552]: value = "task-1240132" [ 1099.315055] env[62552]: _type = "Task" [ 1099.315055] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.324617] env[62552]: DEBUG oslo_vmware.api [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240132, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.430178] env[62552]: INFO nova.compute.manager [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Rescuing [ 1099.430570] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.430803] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquired lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.431060] env[62552]: DEBUG nova.network.neutron [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1099.432459] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1b9133d3-3f5b-427f-a447-a3a720c13ace tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "ed563ce3-40c5-4be0-8948-4d138e81937a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.330s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.699700] env[62552]: DEBUG oslo_concurrency.lockutils [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.688s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.726597] env[62552]: INFO nova.scheduler.client.report [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted allocations for instance 462a3224-d843-45f1-a2bf-69afc18e2572 [ 1099.755940] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: d0aa144e-9bcf-4faa-8d4c-7743118fbe69] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1099.767741] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240131, 'name': Rename_Task, 'duration_secs': 0.205608} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.768012] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1099.768284] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d08a527f-bc65-44af-8616-cedeae5ccf62 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.776153] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1099.776153] env[62552]: value = "task-1240133" [ 1099.776153] env[62552]: _type = "Task" [ 1099.776153] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.786080] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240133, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.520270] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 505d7f8e-c27f-487e-98d8-c840a526d40e] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1100.522076] env[62552]: DEBUG oslo_vmware.api [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240132, 'name': ReconfigVM_Task, 'duration_secs': 0.145994} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.522490] env[62552]: DEBUG oslo_concurrency.lockutils [None req-324df7a7-18a3-4bf9-b068-14fad3cd518a tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "462a3224-d843-45f1-a2bf-69afc18e2572" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.820s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.526040] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267530', 'volume_id': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'name': 'volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '765bbf00-2569-4c49-9634-81d10f375175', 'attached_at': '', 'detached_at': '', 'volume_id': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'serial': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1100.536202] env[62552]: DEBUG oslo_vmware.api [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240133, 'name': PowerOnVM_Task, 'duration_secs': 0.436225} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.536519] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1100.634472] env[62552]: DEBUG nova.compute.manager [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1100.635664] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f697d2-2121-4bce-bc38-583e71747614 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.725050] env[62552]: DEBUG nova.network.neutron [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updating instance_info_cache with network_info: [{"id": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "address": "fa:16:3e:42:30:1b", "network": {"id": "4249dd40-74ec-4271-8d53-48d956819e64", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-928492516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1223fc0363644d1b964ca7f581e9c771", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb18870e-f482-4c7b-8cd4-5c933d3ad294", "external-id": "nsx-vlan-transportzone-76", "segmentation_id": 76, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00e0a1b2-3f", "ovs_interfaceid": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.024415] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 304ad9f2-1965-4a70-aec9-0d2c931b8d95] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1101.154798] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f3b9cd86-4725-43cf-8ad6-8ec3ad89050a tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.407s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.228320] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Releasing lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.527199] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 2c8a5f30-b54f-42ff-a5e0-04eceb1ac734] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1101.567813] env[62552]: DEBUG nova.objects.instance [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid 765bbf00-2569-4c49-9634-81d10f375175 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.766006] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.766565] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.938279] env[62552]: DEBUG nova.compute.manager [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Stashing vm_state: active {{(pid=62552) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 1102.030172] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: ee8d4ad2-a1ed-4a5b-8a02-68d457e17d71] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1102.072743] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6b29b23e-1606-4939-89ee-fcf9392484b4 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.507s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.259997] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.260383] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6266278-ca9c-46ad-9f15-a4fc9cd0904a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.268324] env[62552]: DEBUG nova.compute.manager [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1102.271054] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1102.271054] env[62552]: value = "task-1240134" [ 1102.271054] env[62552]: _type = "Task" [ 1102.271054] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.281401] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240134, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.459572] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.459864] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.533261] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6a4573f0-7152-4462-bb72-58a45b2cbd97] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1102.703387] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.703813] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.786024] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240134, 'name': PowerOffVM_Task, 'duration_secs': 0.204711} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.786024] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1102.786564] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab59c639-510e-4196-bb4d-a4a70efc743a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.805611] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.806414] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c553a285-a692-4b32-ad31-063b2e502b19 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.837313] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.838052] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-537a492e-e28a-4f02-9b93-3b9896673afc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.846170] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1102.846170] env[62552]: value = "task-1240135" [ 1102.846170] env[62552]: _type = "Task" [ 1102.846170] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.854914] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.965096] env[62552]: INFO nova.compute.claims [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1103.036643] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 79166f2c-f864-4d8a-b1dc-e176710400cb] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1103.207269] env[62552]: DEBUG nova.compute.utils [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1103.356730] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] VM already powered off {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1103.356955] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1103.357241] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.357399] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.357579] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1103.357825] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-868da3e2-debd-4103-bbed-fba97aeb3148 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.371502] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1103.371675] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1103.372359] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24660e3a-5515-4cd7-a323-d3b6d5e32a27 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.377377] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1103.377377] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52979953-6704-73b1-cbae-dff08676f7ad" [ 1103.377377] env[62552]: _type = "Task" [ 1103.377377] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.384364] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52979953-6704-73b1-cbae-dff08676f7ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.471790] env[62552]: INFO nova.compute.resource_tracker [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating resource usage from migration 4e07bb93-6387-41f5-a5d0-9ff74a77f96f [ 1103.542182] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 04186fde-a9ed-415a-bd40-312f0347fcc7] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1103.574973] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fca0be1-e456-4199-addb-e69f91d3f3f7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.583701] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f710afa-ae6b-49ac-9eea-652e1ebcf23f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.612718] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0223264d-90fb-440a-a640-0ccfd7219419 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.619698] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f007a23-71cd-46d2-8ec6-0ac6486ae014 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.633246] env[62552]: DEBUG nova.compute.provider_tree [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.710768] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.888769] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52979953-6704-73b1-cbae-dff08676f7ad, 'name': SearchDatastore_Task, 'duration_secs': 0.043146} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.889507] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdf86346-ea82-4d3f-bb37-4dd5bfdd7944 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.895035] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1103.895035] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52db15b6-cb9d-fbe2-e362-0b972291fce7" [ 1103.895035] env[62552]: _type = "Task" [ 1103.895035] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.904275] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52db15b6-cb9d-fbe2-e362-0b972291fce7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.045815] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: c5ed17ee-7421-49d7-975f-1515de1e4b01] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1104.136068] env[62552]: DEBUG nova.scheduler.client.report [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1104.406700] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52db15b6-cb9d-fbe2-e362-0b972291fce7, 'name': SearchDatastore_Task, 'duration_secs': 0.019243} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.406980] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.407267] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] ed563ce3-40c5-4be0-8948-4d138e81937a/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. {{(pid=62552) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1104.407533] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e5f63fb-b1c6-4aec-b4e3-554e78241d78 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.414906] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1104.414906] env[62552]: value = "task-1240136" [ 1104.414906] env[62552]: _type = "Task" [ 1104.414906] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.422493] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240136, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.549452] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: ae517e5f-0cd1-457c-99a6-6cd46fdd6b06] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1104.641029] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.181s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.641237] env[62552]: INFO nova.compute.manager [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Migrating [ 1104.648280] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.843s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.649906] env[62552]: INFO nova.compute.claims [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1104.769023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.769333] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.769580] env[62552]: INFO nova.compute.manager [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Attaching volume 6ecdf541-b516-4d9d-9334-1a32894e38ff to /dev/sdc [ 1104.807155] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6491d7-a242-4ca5-af10-c414f158e9bd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.815169] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5265589f-d406-4dce-a392-9addfd89baa3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.832323] env[62552]: DEBUG nova.virt.block_device [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Updating existing volume attachment record: 00d59864-df4e-4c5c-bdb9-606a219079e9 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1104.926048] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240136, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.425808} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.926406] env[62552]: INFO nova.virt.vmwareapi.ds_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore1] ed563ce3-40c5-4be0-8948-4d138e81937a/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk. [ 1104.927143] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e41a047-81b5-4201-b2e0-2e9a10580aa6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.951785] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] ed563ce3-40c5-4be0-8948-4d138e81937a/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.952142] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65ca65a3-2194-4be9-a59b-d0a1bbd24b11 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.971563] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1104.971563] env[62552]: value = "task-1240137" [ 1104.971563] env[62552]: _type = "Task" [ 1104.971563] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.979865] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240137, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.052781] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 0080a6b3-cac0-44f3-a037-3f94f1daa275] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1105.162667] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.162844] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.162898] env[62552]: DEBUG nova.network.neutron [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1105.481531] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240137, 'name': ReconfigVM_Task, 'duration_secs': 0.374336} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.481922] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Reconfigured VM instance instance-0000006a to attach disk [datastore1] ed563ce3-40c5-4be0-8948-4d138e81937a/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee-rescue.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.482786] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24dd9660-b9ca-43f0-a749-e673905d37fb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.507154] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8d92648-8517-4e35-a267-f3ce39f1e31b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.522677] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1105.522677] env[62552]: value = "task-1240139" [ 1105.522677] env[62552]: _type = "Task" [ 1105.522677] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.530338] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240139, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.557053] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: fbbc360d-9cfd-48f1-80b2-26da2c72c002] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1105.855189] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92067805-7591-447f-adbb-cfa4461faa4b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.863176] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169bbec3-d1b3-4b1c-8cdd-28d190e1eadc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.894449] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b516e33d-0b7b-4308-96ba-d76a63037895 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.902789] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b7069d-e4f0-486b-b6e5-7255362a7c7d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.916991] env[62552]: DEBUG nova.compute.provider_tree [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.033499] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240139, 'name': ReconfigVM_Task, 'duration_secs': 0.152543} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.034076] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1106.034076] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c95deb4-cd43-4afc-80f3-2716f9f8c1c3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.040947] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1106.040947] env[62552]: value = "task-1240140" [ 1106.040947] env[62552]: _type = "Task" [ 1106.040947] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.048577] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240140, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.060927] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 43d1e732-7e9b-4f9f-b67c-f0ef2be91902] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1106.089036] env[62552]: DEBUG nova.network.neutron [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance_info_cache with network_info: [{"id": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "address": "fa:16:3e:9c:12:00", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f6cbf66-fc", "ovs_interfaceid": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.420819] env[62552]: DEBUG nova.scheduler.client.report [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1106.551113] env[62552]: DEBUG oslo_vmware.api [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240140, 'name': PowerOnVM_Task, 'duration_secs': 0.383277} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.551431] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1106.554179] env[62552]: DEBUG nova.compute.manager [None req-3216b30e-920d-4c3d-9bb2-3b6ffcb3ab3a tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1106.554914] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0234e41c-2305-44a1-9956-e65a21ca05a8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.564232] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6c8b5367-4fa0-479b-9382-ff261201e3ef] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1106.591417] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.926626] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.927200] env[62552]: DEBUG nova.compute.manager [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1107.068713] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 11831580-1b58-476a-91ce-a4e55947fd91] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1107.369776] env[62552]: DEBUG nova.compute.manager [req-fbaf45d0-4886-4df6-82cc-1759d7de12ab req-0487463d-6cf0-4c5b-9926-c0ca74e198df service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Received event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1107.369984] env[62552]: DEBUG nova.compute.manager [req-fbaf45d0-4886-4df6-82cc-1759d7de12ab req-0487463d-6cf0-4c5b-9926-c0ca74e198df service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing instance network info cache due to event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1107.370225] env[62552]: DEBUG oslo_concurrency.lockutils [req-fbaf45d0-4886-4df6-82cc-1759d7de12ab req-0487463d-6cf0-4c5b-9926-c0ca74e198df service nova] Acquiring lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.370382] env[62552]: DEBUG oslo_concurrency.lockutils [req-fbaf45d0-4886-4df6-82cc-1759d7de12ab req-0487463d-6cf0-4c5b-9926-c0ca74e198df service nova] Acquired lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.370547] env[62552]: DEBUG nova.network.neutron [req-fbaf45d0-4886-4df6-82cc-1759d7de12ab req-0487463d-6cf0-4c5b-9926-c0ca74e198df service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1107.431865] env[62552]: DEBUG nova.compute.utils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1107.433228] env[62552]: DEBUG nova.compute.manager [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1107.433413] env[62552]: DEBUG nova.network.neutron [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1107.471646] env[62552]: DEBUG nova.policy [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb432303c2394f8795672ceaf02f28aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a0bc011d6794602b2bbe1fc01e4c8b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1107.574707] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: a3ebebf2-c298-4a6e-bcc7-1fe024d68ff0] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1107.717313] env[62552]: DEBUG nova.network.neutron [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Successfully created port: ae5599ec-5b26-49c6-86d1-165c7baf1b2f {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1107.936968] env[62552]: DEBUG nova.compute.manager [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1108.078304] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 526d0f25-3e99-4558-94c0-754ec2a80bad] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1108.108688] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a02d2ae4-4262-499a-aa32-dd717165df86 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.129771] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance '68cb79de-e08c-4ceb-acca-e45952f67248' progress to 0 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1108.184744] env[62552]: DEBUG nova.network.neutron [req-fbaf45d0-4886-4df6-82cc-1759d7de12ab req-0487463d-6cf0-4c5b-9926-c0ca74e198df service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updated VIF entry in instance network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1108.185153] env[62552]: DEBUG nova.network.neutron [req-fbaf45d0-4886-4df6-82cc-1759d7de12ab req-0487463d-6cf0-4c5b-9926-c0ca74e198df service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updating instance_info_cache with network_info: [{"id": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "address": "fa:16:3e:42:30:1b", "network": {"id": "4249dd40-74ec-4271-8d53-48d956819e64", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-928492516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1223fc0363644d1b964ca7f581e9c771", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb18870e-f482-4c7b-8cd4-5c933d3ad294", "external-id": "nsx-vlan-transportzone-76", "segmentation_id": 76, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00e0a1b2-3f", "ovs_interfaceid": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.581556] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 5cee4242-49ae-4cb7-a208-e2982f52fbad] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1108.635260] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1108.635627] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c64593a-4dc0-4cb0-ad16-e70ef4309a6a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.644048] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1108.644048] env[62552]: value = "task-1240142" [ 1108.644048] env[62552]: _type = "Task" [ 1108.644048] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.655016] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.688649] env[62552]: DEBUG oslo_concurrency.lockutils [req-fbaf45d0-4886-4df6-82cc-1759d7de12ab req-0487463d-6cf0-4c5b-9926-c0ca74e198df service nova] Releasing lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.947446] env[62552]: DEBUG nova.compute.manager [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1108.980306] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1108.980306] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1108.980306] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1108.980306] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1108.980577] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.980685] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1108.980894] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1108.981034] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1108.981475] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1108.981475] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1108.982548] env[62552]: DEBUG nova.virt.hardware [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1108.982548] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75d4e95-8457-4a79-a3b7-be8383f5d4c4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.992786] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855d7f46-f422-461a-9c69-754bf6567135 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.086718] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: fb47f136-7859-4756-bc92-fbcf430b4c8b] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1109.140110] env[62552]: DEBUG nova.compute.manager [req-1d503fdd-a96d-4c06-9fdb-f95759329737 req-bac6ecb3-aec0-4052-a745-ef0f8496c635 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Received event network-vif-plugged-ae5599ec-5b26-49c6-86d1-165c7baf1b2f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1109.140346] env[62552]: DEBUG oslo_concurrency.lockutils [req-1d503fdd-a96d-4c06-9fdb-f95759329737 req-bac6ecb3-aec0-4052-a745-ef0f8496c635 service nova] Acquiring lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.140582] env[62552]: DEBUG oslo_concurrency.lockutils [req-1d503fdd-a96d-4c06-9fdb-f95759329737 req-bac6ecb3-aec0-4052-a745-ef0f8496c635 service nova] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.140762] env[62552]: DEBUG oslo_concurrency.lockutils [req-1d503fdd-a96d-4c06-9fdb-f95759329737 req-bac6ecb3-aec0-4052-a745-ef0f8496c635 service nova] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.140932] env[62552]: DEBUG nova.compute.manager [req-1d503fdd-a96d-4c06-9fdb-f95759329737 req-bac6ecb3-aec0-4052-a745-ef0f8496c635 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] No waiting events found dispatching network-vif-plugged-ae5599ec-5b26-49c6-86d1-165c7baf1b2f {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1109.141126] env[62552]: WARNING nova.compute.manager [req-1d503fdd-a96d-4c06-9fdb-f95759329737 req-bac6ecb3-aec0-4052-a745-ef0f8496c635 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Received unexpected event network-vif-plugged-ae5599ec-5b26-49c6-86d1-165c7baf1b2f for instance with vm_state building and task_state spawning. [ 1109.154660] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240142, 'name': PowerOffVM_Task, 'duration_secs': 0.19121} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.154928] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1109.155135] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance '68cb79de-e08c-4ceb-acca-e45952f67248' progress to 17 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1109.225008] env[62552]: DEBUG nova.network.neutron [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Successfully updated port: ae5599ec-5b26-49c6-86d1-165c7baf1b2f {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1109.377430] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1109.377656] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267531', 'volume_id': '6ecdf541-b516-4d9d-9334-1a32894e38ff', 'name': 'volume-6ecdf541-b516-4d9d-9334-1a32894e38ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '765bbf00-2569-4c49-9634-81d10f375175', 'attached_at': '', 'detached_at': '', 'volume_id': '6ecdf541-b516-4d9d-9334-1a32894e38ff', 'serial': '6ecdf541-b516-4d9d-9334-1a32894e38ff'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1109.378583] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8101c3e-fbfa-478f-b07b-c8de8b97548c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.396343] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5eebf8-9dc6-42ad-b7c7-4336363c9d2f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.402016] env[62552]: DEBUG nova.compute.manager [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Received event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1109.402016] env[62552]: DEBUG nova.compute.manager [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing instance network info cache due to event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1109.402016] env[62552]: DEBUG oslo_concurrency.lockutils [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] Acquiring lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.402016] env[62552]: DEBUG oslo_concurrency.lockutils [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] Acquired lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.402016] env[62552]: DEBUG nova.network.neutron [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1109.429510] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] volume-6ecdf541-b516-4d9d-9334-1a32894e38ff/volume-6ecdf541-b516-4d9d-9334-1a32894e38ff.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1109.430427] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3bc2756-e4c4-48cb-a900-be0977827456 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.450668] env[62552]: DEBUG oslo_vmware.api [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1109.450668] env[62552]: value = "task-1240143" [ 1109.450668] env[62552]: _type = "Task" [ 1109.450668] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.462017] env[62552]: DEBUG oslo_vmware.api [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240143, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.591918] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: fea83dde-3181-49dd-a000-5e5cffc8de95] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1109.661833] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1109.662103] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1109.662272] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1109.662745] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1109.662745] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1109.662745] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1109.662938] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1109.663112] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1109.663350] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1109.663530] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1109.663706] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1109.670237] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c749936-23fc-401f-bb47-cf424ccb31d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.688442] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1109.688442] env[62552]: value = "task-1240144" [ 1109.688442] env[62552]: _type = "Task" [ 1109.688442] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.697567] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240144, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.727708] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.727856] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.728020] env[62552]: DEBUG nova.network.neutron [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1109.961315] env[62552]: DEBUG oslo_vmware.api [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240143, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.095667] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: fb881c34-dccc-4703-af8e-c75caafd9b08] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1110.123554] env[62552]: DEBUG nova.network.neutron [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updated VIF entry in instance network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1110.123916] env[62552]: DEBUG nova.network.neutron [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updating instance_info_cache with network_info: [{"id": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "address": "fa:16:3e:42:30:1b", "network": {"id": "4249dd40-74ec-4271-8d53-48d956819e64", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-928492516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1223fc0363644d1b964ca7f581e9c771", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb18870e-f482-4c7b-8cd4-5c933d3ad294", "external-id": "nsx-vlan-transportzone-76", "segmentation_id": 76, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00e0a1b2-3f", "ovs_interfaceid": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.201030] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240144, 'name': ReconfigVM_Task, 'duration_secs': 0.21066} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.201030] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance '68cb79de-e08c-4ceb-acca-e45952f67248' progress to 33 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1110.263178] env[62552]: DEBUG nova.network.neutron [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1110.404300] env[62552]: DEBUG nova.network.neutron [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Updating instance_info_cache with network_info: [{"id": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "address": "fa:16:3e:25:df:0e", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae5599ec-5b", "ovs_interfaceid": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.460469] env[62552]: DEBUG oslo_vmware.api [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240143, 'name': ReconfigVM_Task, 'duration_secs': 0.682313} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.460752] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfigured VM instance instance-00000065 to attach disk [datastore2] volume-6ecdf541-b516-4d9d-9334-1a32894e38ff/volume-6ecdf541-b516-4d9d-9334-1a32894e38ff.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.465352] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dc8f811-9251-43d9-a908-887d41762088 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.481895] env[62552]: DEBUG oslo_vmware.api [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1110.481895] env[62552]: value = "task-1240145" [ 1110.481895] env[62552]: _type = "Task" [ 1110.481895] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.490287] env[62552]: DEBUG oslo_vmware.api [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240145, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.600563] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 1cfeedac-f71e-42e4-a04f-8a0462c85907] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1110.627210] env[62552]: DEBUG oslo_concurrency.lockutils [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] Releasing lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.627488] env[62552]: DEBUG nova.compute.manager [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Received event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1110.627670] env[62552]: DEBUG nova.compute.manager [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing instance network info cache due to event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1110.627887] env[62552]: DEBUG oslo_concurrency.lockutils [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] Acquiring lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.628039] env[62552]: DEBUG oslo_concurrency.lockutils [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] Acquired lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.628212] env[62552]: DEBUG nova.network.neutron [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1110.705799] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1110.706068] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1110.706239] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1110.706432] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1110.706580] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1110.706730] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1110.706933] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1110.707107] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1110.707278] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1110.707446] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1110.707619] env[62552]: DEBUG nova.virt.hardware [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1110.712797] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1110.713083] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25cfcec9-7fe8-475f-bb84-de8c8557384c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.731499] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1110.731499] env[62552]: value = "task-1240146" [ 1110.731499] env[62552]: _type = "Task" [ 1110.731499] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.739409] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240146, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.876694] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "ed563ce3-40c5-4be0-8948-4d138e81937a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.876984] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "ed563ce3-40c5-4be0-8948-4d138e81937a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.877226] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "ed563ce3-40c5-4be0-8948-4d138e81937a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.877421] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "ed563ce3-40c5-4be0-8948-4d138e81937a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.877596] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "ed563ce3-40c5-4be0-8948-4d138e81937a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.879715] env[62552]: INFO nova.compute.manager [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Terminating instance [ 1110.908061] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.908061] env[62552]: DEBUG nova.compute.manager [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Instance network_info: |[{"id": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "address": "fa:16:3e:25:df:0e", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae5599ec-5b", "ovs_interfaceid": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1110.908850] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:df:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae5599ec-5b26-49c6-86d1-165c7baf1b2f', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1110.916610] env[62552]: DEBUG oslo.service.loopingcall [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.916815] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1110.917058] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cd0bfd5-30a4-4475-9ca9-0726128754ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.936625] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1110.936625] env[62552]: value = "task-1240147" [ 1110.936625] env[62552]: _type = "Task" [ 1110.936625] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.944259] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240147, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.996095] env[62552]: DEBUG oslo_vmware.api [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240145, 'name': ReconfigVM_Task, 'duration_secs': 0.150537} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.996297] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267531', 'volume_id': '6ecdf541-b516-4d9d-9334-1a32894e38ff', 'name': 'volume-6ecdf541-b516-4d9d-9334-1a32894e38ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '765bbf00-2569-4c49-9634-81d10f375175', 'attached_at': '', 'detached_at': '', 'volume_id': '6ecdf541-b516-4d9d-9334-1a32894e38ff', 'serial': '6ecdf541-b516-4d9d-9334-1a32894e38ff'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1111.104959] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: e31e3077-04e1-4adb-a0cc-44cf84a89eda] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1111.171324] env[62552]: DEBUG nova.compute.manager [req-f70e2373-6e36-472d-bec8-c98bfc6a8914 req-8a6121cc-c71e-4a0b-b0a6-c180cb4afc5f service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Received event network-changed-ae5599ec-5b26-49c6-86d1-165c7baf1b2f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1111.172325] env[62552]: DEBUG nova.compute.manager [req-f70e2373-6e36-472d-bec8-c98bfc6a8914 req-8a6121cc-c71e-4a0b-b0a6-c180cb4afc5f service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Refreshing instance network info cache due to event network-changed-ae5599ec-5b26-49c6-86d1-165c7baf1b2f. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1111.172325] env[62552]: DEBUG oslo_concurrency.lockutils [req-f70e2373-6e36-472d-bec8-c98bfc6a8914 req-8a6121cc-c71e-4a0b-b0a6-c180cb4afc5f service nova] Acquiring lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.172325] env[62552]: DEBUG oslo_concurrency.lockutils [req-f70e2373-6e36-472d-bec8-c98bfc6a8914 req-8a6121cc-c71e-4a0b-b0a6-c180cb4afc5f service nova] Acquired lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.172557] env[62552]: DEBUG nova.network.neutron [req-f70e2373-6e36-472d-bec8-c98bfc6a8914 req-8a6121cc-c71e-4a0b-b0a6-c180cb4afc5f service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Refreshing network info cache for port ae5599ec-5b26-49c6-86d1-165c7baf1b2f {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1111.246833] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240146, 'name': ReconfigVM_Task, 'duration_secs': 0.209743} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.247306] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1111.248454] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45be52a-7c5b-4fea-9f05-c9f48e41aa53 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.271915] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 68cb79de-e08c-4ceb-acca-e45952f67248/68cb79de-e08c-4ceb-acca-e45952f67248.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.274492] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7d4792b-b8d0-4416-b6bb-24168c9a8dc6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.294075] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1111.294075] env[62552]: value = "task-1240148" [ 1111.294075] env[62552]: _type = "Task" [ 1111.294075] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.302801] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240148, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.383662] env[62552]: DEBUG nova.compute.manager [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1111.383937] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1111.384933] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08368e0-49f9-47aa-8357-ac075a185e9e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.390128] env[62552]: DEBUG nova.network.neutron [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updated VIF entry in instance network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1111.390540] env[62552]: DEBUG nova.network.neutron [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updating instance_info_cache with network_info: [{"id": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "address": "fa:16:3e:42:30:1b", "network": {"id": "4249dd40-74ec-4271-8d53-48d956819e64", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-928492516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1223fc0363644d1b964ca7f581e9c771", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb18870e-f482-4c7b-8cd4-5c933d3ad294", "external-id": "nsx-vlan-transportzone-76", "segmentation_id": 76, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00e0a1b2-3f", "ovs_interfaceid": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.393933] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.394364] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef57a7b5-a52d-496e-9dc4-8b83bea29aa5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.401240] env[62552]: DEBUG oslo_vmware.api [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1111.401240] env[62552]: value = "task-1240149" [ 1111.401240] env[62552]: _type = "Task" [ 1111.401240] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.411304] env[62552]: DEBUG oslo_vmware.api [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.439769] env[62552]: DEBUG nova.compute.manager [req-a5900420-d277-48d2-a668-bc2ebccaef2c req-e86fb36f-a734-40b3-be0d-99b6de2dcad5 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Received event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1111.440153] env[62552]: DEBUG nova.compute.manager [req-a5900420-d277-48d2-a668-bc2ebccaef2c req-e86fb36f-a734-40b3-be0d-99b6de2dcad5 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing instance network info cache due to event network-changed-00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1111.440407] env[62552]: DEBUG oslo_concurrency.lockutils [req-a5900420-d277-48d2-a668-bc2ebccaef2c req-e86fb36f-a734-40b3-be0d-99b6de2dcad5 service nova] Acquiring lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.450612] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240147, 'name': CreateVM_Task, 'duration_secs': 0.410358} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.450767] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1111.451640] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.451831] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.452204] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1111.452699] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b48f4edf-88b7-4dca-a766-049155d00d92 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.457910] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1111.457910] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a4c0ad-0e9c-a556-21ba-79c25b729e0c" [ 1111.457910] env[62552]: _type = "Task" [ 1111.457910] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.467720] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a4c0ad-0e9c-a556-21ba-79c25b729e0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.609143] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 4083ee43-ecea-4ea5-8923-42b348893824] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1111.805367] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240148, 'name': ReconfigVM_Task, 'duration_secs': 0.308744} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.805652] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 68cb79de-e08c-4ceb-acca-e45952f67248/68cb79de-e08c-4ceb-acca-e45952f67248.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1111.805924] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance '68cb79de-e08c-4ceb-acca-e45952f67248' progress to 50 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1111.865944] env[62552]: DEBUG nova.network.neutron [req-f70e2373-6e36-472d-bec8-c98bfc6a8914 req-8a6121cc-c71e-4a0b-b0a6-c180cb4afc5f service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Updated VIF entry in instance network info cache for port ae5599ec-5b26-49c6-86d1-165c7baf1b2f. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1111.866374] env[62552]: DEBUG nova.network.neutron [req-f70e2373-6e36-472d-bec8-c98bfc6a8914 req-8a6121cc-c71e-4a0b-b0a6-c180cb4afc5f service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Updating instance_info_cache with network_info: [{"id": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "address": "fa:16:3e:25:df:0e", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae5599ec-5b", "ovs_interfaceid": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.894942] env[62552]: DEBUG oslo_concurrency.lockutils [req-f173d230-ca49-44a8-8bf5-916745e081d1 req-e49a8d53-39e3-4e4c-92eb-cbab78083ca4 service nova] Releasing lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.895365] env[62552]: DEBUG oslo_concurrency.lockutils [req-a5900420-d277-48d2-a668-bc2ebccaef2c req-e86fb36f-a734-40b3-be0d-99b6de2dcad5 service nova] Acquired lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.895558] env[62552]: DEBUG nova.network.neutron [req-a5900420-d277-48d2-a668-bc2ebccaef2c req-e86fb36f-a734-40b3-be0d-99b6de2dcad5 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Refreshing network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1111.911458] env[62552]: DEBUG oslo_vmware.api [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240149, 'name': PowerOffVM_Task, 'duration_secs': 0.218524} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.911693] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1111.911864] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1111.912117] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50be6bab-dd7b-4427-9de8-4013e827c767 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.968618] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a4c0ad-0e9c-a556-21ba-79c25b729e0c, 'name': SearchDatastore_Task, 'duration_secs': 0.010781} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.968929] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.969224] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1111.969490] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.969616] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.969806] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1111.970083] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47b1c954-11ca-4e52-83b0-a87280fc4251 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.976369] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1111.976539] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1111.976724] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Deleting the datastore file [datastore1] ed563ce3-40c5-4be0-8948-4d138e81937a {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1111.976965] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e3e111c-c236-491e-9d58-eee25006c73b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.981343] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1111.981525] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1111.982236] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60b77f1d-16ac-492f-bb40-aa664d9aee4c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.985919] env[62552]: DEBUG oslo_vmware.api [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for the task: (returnval){ [ 1111.985919] env[62552]: value = "task-1240151" [ 1111.985919] env[62552]: _type = "Task" [ 1111.985919] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.990520] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1111.990520] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5224efe7-9c1b-2205-b42b-d63d958dd292" [ 1111.990520] env[62552]: _type = "Task" [ 1111.990520] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.996751] env[62552]: DEBUG oslo_vmware.api [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.001499] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5224efe7-9c1b-2205-b42b-d63d958dd292, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.034658] env[62552]: DEBUG nova.objects.instance [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid 765bbf00-2569-4c49-9634-81d10f375175 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.112368] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: d8891025-5bdf-4dc9-a2b0-c86f94582ac6] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1112.312137] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbddb6af-3fd6-49f7-9fd0-5ed74e594a92 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.332916] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad31743-14fd-4415-b3c6-c8f1c9b40d8f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.350860] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance '68cb79de-e08c-4ceb-acca-e45952f67248' progress to 67 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1112.368336] env[62552]: DEBUG oslo_concurrency.lockutils [req-f70e2373-6e36-472d-bec8-c98bfc6a8914 req-8a6121cc-c71e-4a0b-b0a6-c180cb4afc5f service nova] Releasing lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.499575] env[62552]: DEBUG oslo_vmware.api [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Task: {'id': task-1240151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215845} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.500230] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1112.500436] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1112.500618] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1112.501145] env[62552]: INFO nova.compute.manager [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1112.501145] env[62552]: DEBUG oslo.service.loopingcall [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1112.501489] env[62552]: DEBUG nova.compute.manager [-] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1112.501612] env[62552]: DEBUG nova.network.neutron [-] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1112.506086] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5224efe7-9c1b-2205-b42b-d63d958dd292, 'name': SearchDatastore_Task, 'duration_secs': 0.018296} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.507177] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6c061c9-4da8-4c13-950c-e771bfe7ddf7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.512025] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1112.512025] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a4e96a-0495-26ba-17b2-3dcc809511dc" [ 1112.512025] env[62552]: _type = "Task" [ 1112.512025] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.522060] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a4e96a-0495-26ba-17b2-3dcc809511dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.539698] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c696408f-83e6-4a49-8947-b01d1a0259ba tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.770s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.583440] env[62552]: DEBUG nova.network.neutron [req-a5900420-d277-48d2-a668-bc2ebccaef2c req-e86fb36f-a734-40b3-be0d-99b6de2dcad5 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updated VIF entry in instance network info cache for port 00e0a1b2-3fba-44a1-a14d-9846dd2d281d. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1112.583871] env[62552]: DEBUG nova.network.neutron [req-a5900420-d277-48d2-a668-bc2ebccaef2c req-e86fb36f-a734-40b3-be0d-99b6de2dcad5 service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updating instance_info_cache with network_info: [{"id": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "address": "fa:16:3e:42:30:1b", "network": {"id": "4249dd40-74ec-4271-8d53-48d956819e64", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-928492516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1223fc0363644d1b964ca7f581e9c771", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb18870e-f482-4c7b-8cd4-5c933d3ad294", "external-id": "nsx-vlan-transportzone-76", "segmentation_id": 76, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00e0a1b2-3f", "ovs_interfaceid": "00e0a1b2-3fba-44a1-a14d-9846dd2d281d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.615419] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 22357d4e-9771-477c-9fc3-fe3d76f6e902] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1112.888536] env[62552]: DEBUG nova.network.neutron [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Port 4f6cbf66-fc98-4868-82ef-2f234be029cb binding to destination host cpu-1 is already ACTIVE {{(pid=62552) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1113.026176] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a4e96a-0495-26ba-17b2-3dcc809511dc, 'name': SearchDatastore_Task, 'duration_secs': 0.010998} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.026403] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.026659] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 888ed8fb-18e6-4f69-a27c-bd24c19e0a12/888ed8fb-18e6-4f69-a27c-bd24c19e0a12.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1113.026908] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e25a0aa3-2456-412e-89ce-5166b8988263 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.035206] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1113.035206] env[62552]: value = "task-1240152" [ 1113.035206] env[62552]: _type = "Task" [ 1113.035206] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.042961] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240152, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.086131] env[62552]: DEBUG oslo_concurrency.lockutils [req-a5900420-d277-48d2-a668-bc2ebccaef2c req-e86fb36f-a734-40b3-be0d-99b6de2dcad5 service nova] Releasing lock "refresh_cache-ed563ce3-40c5-4be0-8948-4d138e81937a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.119691] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: c52efdeb-9efc-4ba1-b9af-4dd0a925517f] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1113.251174] env[62552]: DEBUG nova.network.neutron [-] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.351242] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.351544] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.466798] env[62552]: DEBUG nova.compute.manager [req-317b3872-7523-44cc-afb5-40e0b1167cc9 req-22cd5060-13aa-4dfb-844e-c5623586b04f service nova] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Received event network-vif-deleted-00e0a1b2-3fba-44a1-a14d-9846dd2d281d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1113.545517] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240152, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448419} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.545800] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 888ed8fb-18e6-4f69-a27c-bd24c19e0a12/888ed8fb-18e6-4f69-a27c-bd24c19e0a12.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1113.546033] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1113.546363] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-017eaaf4-3d16-4c98-a14d-20fcf908211e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.552379] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1113.552379] env[62552]: value = "task-1240153" [ 1113.552379] env[62552]: _type = "Task" [ 1113.552379] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.560168] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240153, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.622761] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 474dfa9e-d003-478f-b48f-09e0c0452ffd] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1113.753838] env[62552]: INFO nova.compute.manager [-] [instance: ed563ce3-40c5-4be0-8948-4d138e81937a] Took 1.25 seconds to deallocate network for instance. [ 1113.855140] env[62552]: INFO nova.compute.manager [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Detaching volume 65d3ba2b-bf26-458c-a32c-5eaf30bf1a81 [ 1113.888707] env[62552]: INFO nova.virt.block_device [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Attempting to driver detach volume 65d3ba2b-bf26-458c-a32c-5eaf30bf1a81 from mountpoint /dev/sdb [ 1113.889213] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1113.889213] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267530', 'volume_id': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'name': 'volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '765bbf00-2569-4c49-9634-81d10f375175', 'attached_at': '', 'detached_at': '', 'volume_id': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'serial': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1113.889996] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1baedb-67ea-4ffb-9ecf-3f3d7ce5b1ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.924014] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d0d9f0-5506-4bd7-aecc-4cbccb477a6c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.937454] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "68cb79de-e08c-4ceb-acca-e45952f67248-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.937623] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "68cb79de-e08c-4ceb-acca-e45952f67248-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.937789] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "68cb79de-e08c-4ceb-acca-e45952f67248-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.943755] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f9bd7d-e759-4bf7-80e1-a229420455d4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.968344] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9facd92c-0a7f-4aa6-9fa9-ab8030bd5fcf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.984765] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] The volume has not been displaced from its original location: [datastore1] volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81/volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1113.989957] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfiguring VM instance instance-00000065 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1113.990286] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b49a71a-acc2-420e-bfc1-e18b495e30bf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.008343] env[62552]: DEBUG oslo_vmware.api [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1114.008343] env[62552]: value = "task-1240154" [ 1114.008343] env[62552]: _type = "Task" [ 1114.008343] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.015949] env[62552]: DEBUG oslo_vmware.api [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240154, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.061988] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240153, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063196} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.062275] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1114.063050] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3917fa29-f8fb-48df-a7d4-1582fe1bd93a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.084490] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 888ed8fb-18e6-4f69-a27c-bd24c19e0a12/888ed8fb-18e6-4f69-a27c-bd24c19e0a12.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1114.085107] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a1c59b5-7a29-41a5-a404-64eb8c1e2ea9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.105199] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1114.105199] env[62552]: value = "task-1240155" [ 1114.105199] env[62552]: _type = "Task" [ 1114.105199] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.112827] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240155, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.125742] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 88d0f502-fc3f-429e-bdf6-a1ebacec2117] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1114.261163] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.261446] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.261691] env[62552]: DEBUG nova.objects.instance [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lazy-loading 'resources' on Instance uuid ed563ce3-40c5-4be0-8948-4d138e81937a {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1114.518705] env[62552]: DEBUG oslo_vmware.api [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240154, 'name': ReconfigVM_Task, 'duration_secs': 0.2212} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.518989] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfigured VM instance instance-00000065 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1114.523526] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01405bf8-53ac-45b4-84c5-8bf2908831c3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.538788] env[62552]: DEBUG oslo_vmware.api [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1114.538788] env[62552]: value = "task-1240156" [ 1114.538788] env[62552]: _type = "Task" [ 1114.538788] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.546517] env[62552]: DEBUG oslo_vmware.api [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240156, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.615857] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240155, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.629033] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: df784aaf-0d0d-4c7f-bfb1-f7fde9fabc28] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1114.870856] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe90054a-538e-45d7-a586-85fa3679505e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.879091] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e89916-dc56-4765-9241-3e3df449f0dd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.913750] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a9139d-8fcc-4e94-a017-855bd2eac449 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.922211] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cfba5b-32a7-4979-a319-a92f6bb6afea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.938474] env[62552]: DEBUG nova.compute.provider_tree [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.980138] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.980349] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.980533] env[62552]: DEBUG nova.network.neutron [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1115.050044] env[62552]: DEBUG oslo_vmware.api [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240156, 'name': ReconfigVM_Task, 'duration_secs': 0.146836} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.050352] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267530', 'volume_id': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'name': 'volume-65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '765bbf00-2569-4c49-9634-81d10f375175', 'attached_at': '', 'detached_at': '', 'volume_id': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81', 'serial': '65d3ba2b-bf26-458c-a32c-5eaf30bf1a81'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1115.116070] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240155, 'name': ReconfigVM_Task, 'duration_secs': 0.845383} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.116360] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 888ed8fb-18e6-4f69-a27c-bd24c19e0a12/888ed8fb-18e6-4f69-a27c-bd24c19e0a12.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1115.116974] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5447dccb-d200-4a09-a9ed-3acf9b5bbe40 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.123829] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1115.123829] env[62552]: value = "task-1240157" [ 1115.123829] env[62552]: _type = "Task" [ 1115.123829] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.131381] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 80478878-ff82-4ed6-a851-8eb2bec01e22] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1115.133130] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240157, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.441682] env[62552]: DEBUG nova.scheduler.client.report [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1115.602747] env[62552]: DEBUG nova.objects.instance [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid 765bbf00-2569-4c49-9634-81d10f375175 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.633904] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: fef94e30-f946-4d7f-a108-f32cad505bc4] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1115.635681] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240157, 'name': Rename_Task, 'duration_secs': 0.148657} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.636306] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1115.636546] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3079aa35-4441-4268-bbd4-009909d460cb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.643521] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1115.643521] env[62552]: value = "task-1240158" [ 1115.643521] env[62552]: _type = "Task" [ 1115.643521] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.653101] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.787210] env[62552]: DEBUG nova.network.neutron [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance_info_cache with network_info: [{"id": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "address": "fa:16:3e:9c:12:00", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f6cbf66-fc", "ovs_interfaceid": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.946597] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.685s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.975194] env[62552]: INFO nova.scheduler.client.report [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Deleted allocations for instance ed563ce3-40c5-4be0-8948-4d138e81937a [ 1116.140181] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 3b333977-0b4b-4ccf-bed5-231215abae3c] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1116.153460] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240158, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.291561] env[62552]: DEBUG oslo_concurrency.lockutils [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.486975] env[62552]: DEBUG oslo_concurrency.lockutils [None req-b0b6fe8b-69ef-4892-bdca-d888c61b11c2 tempest-ServerRescueTestJSONUnderV235-1463233169 tempest-ServerRescueTestJSONUnderV235-1463233169-project-member] Lock "ed563ce3-40c5-4be0-8948-4d138e81937a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.610s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.611173] env[62552]: DEBUG oslo_concurrency.lockutils [None req-dc7e2cff-7732-4489-b7a6-fc30e680a7a9 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.259s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.640521] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.640780] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.642532] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 36233c83-1170-4a7a-be0b-6a0e8b139de7] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1116.654517] env[62552]: DEBUG oslo_vmware.api [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240158, 'name': PowerOnVM_Task, 'duration_secs': 0.986338} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.654792] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1116.654992] env[62552]: INFO nova.compute.manager [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Took 7.71 seconds to spawn the instance on the hypervisor. [ 1116.655189] env[62552]: DEBUG nova.compute.manager [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1116.655952] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591d31b7-df1d-423b-9323-21a9644521d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.848056] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d32ef4c-ff20-4b08-aeaf-d326270de0b8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.869083] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4275aab7-e053-4f18-a1f6-628c0ba66d16 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.876082] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance '68cb79de-e08c-4ceb-acca-e45952f67248' progress to 83 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1117.147365] env[62552]: INFO nova.compute.manager [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Detaching volume 6ecdf541-b516-4d9d-9334-1a32894e38ff [ 1117.148915] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 34736dd0-e617-475e-baa2-cb372db1afb2] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1117.175765] env[62552]: INFO nova.compute.manager [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Took 14.40 seconds to build instance. [ 1117.196786] env[62552]: INFO nova.virt.block_device [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Attempting to driver detach volume 6ecdf541-b516-4d9d-9334-1a32894e38ff from mountpoint /dev/sdc [ 1117.197831] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1117.197831] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267531', 'volume_id': '6ecdf541-b516-4d9d-9334-1a32894e38ff', 'name': 'volume-6ecdf541-b516-4d9d-9334-1a32894e38ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '765bbf00-2569-4c49-9634-81d10f375175', 'attached_at': '', 'detached_at': '', 'volume_id': '6ecdf541-b516-4d9d-9334-1a32894e38ff', 'serial': '6ecdf541-b516-4d9d-9334-1a32894e38ff'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1117.198388] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa30078-147d-4891-9e7a-b956e926dfb1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.221632] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc11ebe-77c5-4e2c-bf6b-4af596d1902c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.231283] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a042f49d-2d90-4bca-8d15-53bc72fb9f74 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.253055] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-935000bc-c78c-477b-9cce-1c223c7fa100 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.269735] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] The volume has not been displaced from its original location: [datastore2] volume-6ecdf541-b516-4d9d-9334-1a32894e38ff/volume-6ecdf541-b516-4d9d-9334-1a32894e38ff.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1117.275229] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfiguring VM instance instance-00000065 to detach disk 2002 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1117.275622] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02633820-b099-48bc-8250-93aba1653d18 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.296338] env[62552]: DEBUG oslo_vmware.api [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1117.296338] env[62552]: value = "task-1240159" [ 1117.296338] env[62552]: _type = "Task" [ 1117.296338] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.306067] env[62552]: DEBUG oslo_vmware.api [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240159, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.383853] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1117.384440] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-481e277a-a06d-44f3-9993-e7e4ffaf4521 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.394578] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1117.394578] env[62552]: value = "task-1240160" [ 1117.394578] env[62552]: _type = "Task" [ 1117.394578] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.404936] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240160, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.652731] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 714fb65d-9f80-4a81-a637-3e4398405d9b] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1117.677863] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9474adcb-8b2c-4150-a37a-065a8100d653 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.911s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.810018] env[62552]: DEBUG oslo_vmware.api [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240159, 'name': ReconfigVM_Task, 'duration_secs': 0.34718} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.810018] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Reconfigured VM instance instance-00000065 to detach disk 2002 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1117.812357] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0810599f-ca40-454f-9083-7d2496eef65d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.830240] env[62552]: DEBUG oslo_vmware.api [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1117.830240] env[62552]: value = "task-1240161" [ 1117.830240] env[62552]: _type = "Task" [ 1117.830240] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.840046] env[62552]: DEBUG oslo_vmware.api [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240161, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.905676] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240160, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.157164] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 51bbcc4e-8251-4b38-9d36-8aea8fc7705d] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1118.340886] env[62552]: DEBUG oslo_vmware.api [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240161, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.348712] env[62552]: DEBUG nova.compute.manager [req-60403e7e-349d-4e0a-9d69-566ad342ea49 req-1db30af3-4e9d-43b3-8054-a732204c2ed5 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Received event network-changed-ae5599ec-5b26-49c6-86d1-165c7baf1b2f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1118.348915] env[62552]: DEBUG nova.compute.manager [req-60403e7e-349d-4e0a-9d69-566ad342ea49 req-1db30af3-4e9d-43b3-8054-a732204c2ed5 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Refreshing instance network info cache due to event network-changed-ae5599ec-5b26-49c6-86d1-165c7baf1b2f. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1118.349164] env[62552]: DEBUG oslo_concurrency.lockutils [req-60403e7e-349d-4e0a-9d69-566ad342ea49 req-1db30af3-4e9d-43b3-8054-a732204c2ed5 service nova] Acquiring lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.349321] env[62552]: DEBUG oslo_concurrency.lockutils [req-60403e7e-349d-4e0a-9d69-566ad342ea49 req-1db30af3-4e9d-43b3-8054-a732204c2ed5 service nova] Acquired lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.349488] env[62552]: DEBUG nova.network.neutron [req-60403e7e-349d-4e0a-9d69-566ad342ea49 req-1db30af3-4e9d-43b3-8054-a732204c2ed5 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Refreshing network info cache for port ae5599ec-5b26-49c6-86d1-165c7baf1b2f {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1118.413365] env[62552]: DEBUG oslo_vmware.api [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240160, 'name': PowerOnVM_Task, 'duration_secs': 0.909262} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.413779] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1118.414444] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-a1cb96a7-a795-43c3-a1db-d29c80d62dc2 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance '68cb79de-e08c-4ceb-acca-e45952f67248' progress to 100 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1118.661122] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 4b80bc41-1b63-444e-8039-696e3d78d9a1] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1118.844814] env[62552]: DEBUG oslo_vmware.api [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240161, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.164582] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 00b06df7-e7ae-48fd-8887-65647931d5cf] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1119.344481] env[62552]: DEBUG oslo_vmware.api [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240161, 'name': ReconfigVM_Task, 'duration_secs': 1.269108} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.344794] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267531', 'volume_id': '6ecdf541-b516-4d9d-9334-1a32894e38ff', 'name': 'volume-6ecdf541-b516-4d9d-9334-1a32894e38ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '765bbf00-2569-4c49-9634-81d10f375175', 'attached_at': '', 'detached_at': '', 'volume_id': '6ecdf541-b516-4d9d-9334-1a32894e38ff', 'serial': '6ecdf541-b516-4d9d-9334-1a32894e38ff'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1119.359928] env[62552]: DEBUG nova.network.neutron [req-60403e7e-349d-4e0a-9d69-566ad342ea49 req-1db30af3-4e9d-43b3-8054-a732204c2ed5 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Updated VIF entry in instance network info cache for port ae5599ec-5b26-49c6-86d1-165c7baf1b2f. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1119.360331] env[62552]: DEBUG nova.network.neutron [req-60403e7e-349d-4e0a-9d69-566ad342ea49 req-1db30af3-4e9d-43b3-8054-a732204c2ed5 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Updating instance_info_cache with network_info: [{"id": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "address": "fa:16:3e:25:df:0e", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae5599ec-5b", "ovs_interfaceid": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.668033] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 0418260a-aa27-4955-ab15-b180ec04f0b3] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1119.863589] env[62552]: DEBUG oslo_concurrency.lockutils [req-60403e7e-349d-4e0a-9d69-566ad342ea49 req-1db30af3-4e9d-43b3-8054-a732204c2ed5 service nova] Releasing lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.891294] env[62552]: DEBUG nova.objects.instance [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'flavor' on Instance uuid 765bbf00-2569-4c49-9634-81d10f375175 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.171183] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: ce5d0165-65f1-4505-9c46-1129c56a8913] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1120.674191] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 8707ce4b-677e-4f13-86f8-3e327d19380b] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1120.898989] env[62552]: DEBUG oslo_concurrency.lockutils [None req-82b9cbd4-6a45-4826-964c-8389726d9756 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.258s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.012021] env[62552]: DEBUG nova.network.neutron [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Port 4f6cbf66-fc98-4868-82ef-2f234be029cb binding to destination host cpu-1 is already ACTIVE {{(pid=62552) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1121.012324] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.012584] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.012781] env[62552]: DEBUG nova.network.neutron [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1121.177344] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 1472ca01-ba04-4e7d-a9cf-ccaf29c04282] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1121.680657] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: cdee22a9-4327-47af-ab14-8403a06b2802] Instance has had 0 of 5 cleanup attempts {{(pid=62552) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1121.839855] env[62552]: DEBUG nova.network.neutron [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance_info_cache with network_info: [{"id": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "address": "fa:16:3e:9c:12:00", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f6cbf66-fc", "ovs_interfaceid": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.099554] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.099940] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.100221] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "765bbf00-2569-4c49-9634-81d10f375175-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.100429] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.100690] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.103273] env[62552]: INFO nova.compute.manager [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Terminating instance [ 1122.170005] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "60238708-4d45-4d73-bd8d-2cbc25704259" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.170264] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.184324] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.184652] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Cleaning up deleted instances with incomplete migration {{(pid=62552) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11609}} [ 1122.342518] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.608385] env[62552]: DEBUG nova.compute.manager [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1122.608714] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1122.610012] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ff2b29-8552-4cd5-a1e5-c7bf66c8a18d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.621307] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1122.621728] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e5ffffc-9df7-4c96-a7b6-268a4fec0520 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.628525] env[62552]: DEBUG oslo_vmware.api [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1122.628525] env[62552]: value = "task-1240164" [ 1122.628525] env[62552]: _type = "Task" [ 1122.628525] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.637592] env[62552]: DEBUG oslo_vmware.api [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.674016] env[62552]: DEBUG nova.compute.utils [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1122.687729] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.846328] env[62552]: DEBUG nova.compute.manager [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62552) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1122.846574] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.846939] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.139554] env[62552]: DEBUG oslo_vmware.api [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240164, 'name': PowerOffVM_Task, 'duration_secs': 0.213202} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.139830] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1123.140018] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1123.140280] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4e233a0-d3a6-4f00-9a70-1a06c9e2819c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.177312] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.245807] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1123.246317] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1123.246537] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Deleting the datastore file [datastore2] 765bbf00-2569-4c49-9634-81d10f375175 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1123.246933] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44869938-9cfc-4f8b-8481-4208fb4eedf7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.254831] env[62552]: DEBUG oslo_vmware.api [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for the task: (returnval){ [ 1123.254831] env[62552]: value = "task-1240166" [ 1123.254831] env[62552]: _type = "Task" [ 1123.254831] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.264384] env[62552]: DEBUG oslo_vmware.api [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.350281] env[62552]: DEBUG nova.objects.instance [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'migration_context' on Instance uuid 68cb79de-e08c-4ceb-acca-e45952f67248 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1123.765160] env[62552]: DEBUG oslo_vmware.api [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Task: {'id': task-1240166, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171909} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.765436] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1123.765625] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1123.765863] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1123.765971] env[62552]: INFO nova.compute.manager [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1123.766239] env[62552]: DEBUG oslo.service.loopingcall [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1123.766434] env[62552]: DEBUG nova.compute.manager [-] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1123.766530] env[62552]: DEBUG nova.network.neutron [-] [instance: 765bbf00-2569-4c49-9634-81d10f375175] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1123.962872] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14b16c7-ad68-4bd6-809f-bad693a4f82b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.971068] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fdb980-e06a-4c5e-8e28-e7a835c7cce3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.002839] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e47d3f4-349f-40e1-957b-c04ee14f4835 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.010978] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a047404-bd6a-4340-bc51-67e171086ea2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.026479] env[62552]: DEBUG nova.compute.provider_tree [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.237931] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "60238708-4d45-4d73-bd8d-2cbc25704259" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.238238] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.238484] env[62552]: INFO nova.compute.manager [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Attaching volume 1900f85e-47dc-4137-825d-646342d03f76 to /dev/sdb [ 1124.273088] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae70000-894f-4fb9-8a03-ef40f11f4111 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.281506] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f007023c-64ca-4d4e-b7ad-23c504b60a94 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.299572] env[62552]: DEBUG nova.virt.block_device [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Updating existing volume attachment record: 8667964d-0e9c-4dde-b776-d6ea1cdfd9af {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1124.310483] env[62552]: DEBUG nova.compute.manager [req-ad9310d1-3ed4-4390-ad6e-2a453c375af6 req-e831d61e-6f0c-43ce-8532-1efe455dc620 service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Received event network-vif-deleted-28ef7f98-6d55-4c0f-9af2-e430a81bbec8 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1124.310688] env[62552]: INFO nova.compute.manager [req-ad9310d1-3ed4-4390-ad6e-2a453c375af6 req-e831d61e-6f0c-43ce-8532-1efe455dc620 service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Neutron deleted interface 28ef7f98-6d55-4c0f-9af2-e430a81bbec8; detaching it from the instance and deleting it from the info cache [ 1124.310894] env[62552]: DEBUG nova.network.neutron [req-ad9310d1-3ed4-4390-ad6e-2a453c375af6 req-e831d61e-6f0c-43ce-8532-1efe455dc620 service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.529708] env[62552]: DEBUG nova.scheduler.client.report [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1124.786409] env[62552]: DEBUG nova.network.neutron [-] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.813846] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-14331619-d8c8-44d9-b616-296cbabfcf6b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.824809] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea53d9d8-2182-48a2-96aa-d22eb0116646 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.854654] env[62552]: DEBUG nova.compute.manager [req-ad9310d1-3ed4-4390-ad6e-2a453c375af6 req-e831d61e-6f0c-43ce-8532-1efe455dc620 service nova] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Detach interface failed, port_id=28ef7f98-6d55-4c0f-9af2-e430a81bbec8, reason: Instance 765bbf00-2569-4c49-9634-81d10f375175 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1125.189736] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1125.190678] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1125.289249] env[62552]: INFO nova.compute.manager [-] [instance: 765bbf00-2569-4c49-9634-81d10f375175] Took 1.52 seconds to deallocate network for instance. [ 1125.540843] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.694s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.697047] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1125.697207] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1125.697293] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Rebuilding the list of instances to heal {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1125.795902] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.796206] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.796442] env[62552]: DEBUG nova.objects.instance [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lazy-loading 'resources' on Instance uuid 765bbf00-2569-4c49-9634-81d10f375175 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.243704] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.243948] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquired lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.244149] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Forcefully refreshing network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1126.244316] env[62552]: DEBUG nova.objects.instance [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lazy-loading 'info_cache' on Instance uuid 6a5b5f4e-0c84-447d-a3da-8258512abb7c {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.384845] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b117837d-e856-45d1-9e10-993aee28044f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.392725] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4951f385-1d0b-4a54-b994-e24b6e70e6d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.422975] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba25964-2f47-42fd-a615-a7e6668a6327 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.430559] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ab1f87-f3e8-4083-bb94-f893e296ff9c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.443701] env[62552]: DEBUG nova.compute.provider_tree [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.946834] env[62552]: DEBUG nova.scheduler.client.report [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1127.078838] env[62552]: INFO nova.compute.manager [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Swapping old allocation on dict_keys(['b52e1b51-8a99-4529-bad8-00cd27fc0fb8']) held by migration 4e07bb93-6387-41f5-a5d0-9ff74a77f96f for instance [ 1127.098261] env[62552]: DEBUG nova.scheduler.client.report [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Overwriting current allocation {'allocations': {'b52e1b51-8a99-4529-bad8-00cd27fc0fb8': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 141}}, 'project_id': '8495bac65967441996ecca1fd22a4da4', 'user_id': 'bdea5d6f50634b5e89852744db56db67', 'consumer_generation': 1} on consumer 68cb79de-e08c-4ceb-acca-e45952f67248 {{(pid=62552) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2033}} [ 1127.170557] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.170759] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.170940] env[62552]: DEBUG nova.network.neutron [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1127.452377] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.656s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.468729] env[62552]: INFO nova.scheduler.client.report [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Deleted allocations for instance 765bbf00-2569-4c49-9634-81d10f375175 [ 1127.886914] env[62552]: DEBUG nova.network.neutron [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance_info_cache with network_info: [{"id": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "address": "fa:16:3e:9c:12:00", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f6cbf66-fc", "ovs_interfaceid": "4f6cbf66-fc98-4868-82ef-2f234be029cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.956275] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updating instance_info_cache with network_info: [{"id": "5e69a23a-a63b-4d24-9729-859b537f945b", "address": "fa:16:3e:0b:c1:98", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e69a23a-a6", "ovs_interfaceid": "5e69a23a-a63b-4d24-9729-859b537f945b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.977047] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12535589-f49a-4aa2-ac27-f9a755756e78 tempest-AttachVolumeTestJSON-1474809303 tempest-AttachVolumeTestJSON-1474809303-project-member] Lock "765bbf00-2569-4c49-9634-81d10f375175" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.877s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.389886] env[62552]: DEBUG oslo_concurrency.lockutils [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-68cb79de-e08c-4ceb-acca-e45952f67248" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.390388] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1128.390675] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1552b2d-5fa6-43f7-88e8-1785221c3e39 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.400065] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1128.400065] env[62552]: value = "task-1240171" [ 1128.400065] env[62552]: _type = "Task" [ 1128.400065] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.409916] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240171, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.458644] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Releasing lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.459226] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updated the network info_cache for instance {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1128.459226] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.459226] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.459443] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.459518] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.459658] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.459797] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.459926] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1128.460087] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.846734] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1128.847025] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267534', 'volume_id': '1900f85e-47dc-4137-825d-646342d03f76', 'name': 'volume-1900f85e-47dc-4137-825d-646342d03f76', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60238708-4d45-4d73-bd8d-2cbc25704259', 'attached_at': '', 'detached_at': '', 'volume_id': '1900f85e-47dc-4137-825d-646342d03f76', 'serial': '1900f85e-47dc-4137-825d-646342d03f76'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1128.847913] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb4b722-870f-432e-be0c-111883028302 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.865252] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d533b3-6ea5-4dcc-9829-e172a8d925cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.892677] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] volume-1900f85e-47dc-4137-825d-646342d03f76/volume-1900f85e-47dc-4137-825d-646342d03f76.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.893120] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-065d87ba-4f65-43df-b3ab-3f85335d58c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.915968] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240171, 'name': PowerOffVM_Task, 'duration_secs': 0.159796} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.917307] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1128.918042] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1128.918230] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1128.918407] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1128.918598] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1128.918749] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1128.918902] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1128.919125] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1128.919290] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1128.919459] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1128.919629] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1128.919805] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1128.925517] env[62552]: DEBUG oslo_vmware.api [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1128.925517] env[62552]: value = "task-1240174" [ 1128.925517] env[62552]: _type = "Task" [ 1128.925517] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.925832] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccc94aae-22bf-4274-b946-57346bb63f8a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.948162] env[62552]: DEBUG oslo_vmware.api [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240174, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.949949] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1128.949949] env[62552]: value = "task-1240175" [ 1128.949949] env[62552]: _type = "Task" [ 1128.949949] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.959580] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240175, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.962788] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.963283] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.963404] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.963727] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1128.964681] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64e08a9-6ea7-44b7-828c-dc995e23cd7b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.974248] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79017bd7-78be-4714-ac05-b423e00cc9ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.990222] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925ce284-149e-4b6d-b105-ea7b89a7dc10 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.997402] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e6f9b9-81e9-4cf1-a7da-45e142a11c81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.027953] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179819MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1129.028221] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.028380] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.448358] env[62552]: DEBUG oslo_vmware.api [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240174, 'name': ReconfigVM_Task, 'duration_secs': 0.43634} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.448714] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Reconfigured VM instance instance-00000069 to attach disk [datastore2] volume-1900f85e-47dc-4137-825d-646342d03f76/volume-1900f85e-47dc-4137-825d-646342d03f76.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.453866] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0be26449-73c7-43cb-99db-e00c51b0ad77 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.473181] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240175, 'name': ReconfigVM_Task, 'duration_secs': 0.145424} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.474848] env[62552]: DEBUG oslo_vmware.api [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1129.474848] env[62552]: value = "task-1240176" [ 1129.474848] env[62552]: _type = "Task" [ 1129.474848] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.475733] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1825f5d4-dab4-4869-a0bf-ccdf1030a1b6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.499324] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1129.499735] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1129.500019] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1129.500333] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1129.500606] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1129.500863] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1129.501213] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1129.501518] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1129.501825] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1129.502095] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1129.502412] env[62552]: DEBUG nova.virt.hardware [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.507730] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0239372-6569-4c88-a216-02af536a075b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.510070] env[62552]: DEBUG oslo_vmware.api [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240176, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.513737] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1129.513737] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d04503-d341-0a0b-a84e-0a013a867d23" [ 1129.513737] env[62552]: _type = "Task" [ 1129.513737] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.522014] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d04503-d341-0a0b-a84e-0a013a867d23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.990024] env[62552]: DEBUG oslo_vmware.api [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240176, 'name': ReconfigVM_Task, 'duration_secs': 0.205627} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.990024] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267534', 'volume_id': '1900f85e-47dc-4137-825d-646342d03f76', 'name': 'volume-1900f85e-47dc-4137-825d-646342d03f76', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60238708-4d45-4d73-bd8d-2cbc25704259', 'attached_at': '', 'detached_at': '', 'volume_id': '1900f85e-47dc-4137-825d-646342d03f76', 'serial': '1900f85e-47dc-4137-825d-646342d03f76'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1130.023919] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d04503-d341-0a0b-a84e-0a013a867d23, 'name': SearchDatastore_Task, 'duration_secs': 0.007576} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.029081] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1130.029357] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12937d4c-76ea-4209-90c5-c02756fab45a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.050543] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1130.050543] env[62552]: value = "task-1240177" [ 1130.050543] env[62552]: _type = "Task" [ 1130.050543] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.058584] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240177, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.060616] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6a5b5f4e-0c84-447d-a3da-8258512abb7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1130.060758] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 60238708-4d45-4d73-bd8d-2cbc25704259 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1130.060881] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 158d4227-90ca-41ae-821b-efd353928cb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1130.060998] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 888ed8fb-18e6-4f69-a27c-bd24c19e0a12 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1130.061140] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 68cb79de-e08c-4ceb-acca-e45952f67248 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1130.061317] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1130.061455] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1130.126738] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533bcf1f-50ba-4fed-a728-bf8b4be16419 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.134291] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868b24f5-c8f0-4e0c-907e-8d98147b0750 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.164318] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18967288-b5f0-4809-9d6b-0a6e060ea01a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.171526] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530e960e-07ce-4af3-b615-cee6ff8fa93c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.184332] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.560430] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240177, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.687896] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1131.025510] env[62552]: DEBUG nova.objects.instance [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'flavor' on Instance uuid 60238708-4d45-4d73-bd8d-2cbc25704259 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.060673] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240177, 'name': ReconfigVM_Task, 'duration_secs': 0.713158} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.060978] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1131.061847] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946f20c1-8dec-4547-b7a3-7fc96f9aff67 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.085041] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 68cb79de-e08c-4ceb-acca-e45952f67248/68cb79de-e08c-4ceb-acca-e45952f67248.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1131.085340] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06b98901-903f-491c-bf4d-722d7cb69fac {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.103320] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1131.103320] env[62552]: value = "task-1240179" [ 1131.103320] env[62552]: _type = "Task" [ 1131.103320] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.111886] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240179, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.193084] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1131.193297] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.165s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.530857] env[62552]: DEBUG oslo_concurrency.lockutils [None req-5f39c327-d366-4ed5-a942-990742085b12 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.292s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.613560] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240179, 'name': ReconfigVM_Task, 'duration_secs': 0.258539} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.613919] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 68cb79de-e08c-4ceb-acca-e45952f67248/68cb79de-e08c-4ceb-acca-e45952f67248.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1131.614694] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c061ed38-2459-4106-9015-a09ac0b79754 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.632294] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f006ac9-3340-45d0-aae7-07fd09179363 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.651587] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45ed74e-ce6b-4818-8089-49f6e561e4f5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.668957] env[62552]: DEBUG oslo_concurrency.lockutils [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "60238708-4d45-4d73-bd8d-2cbc25704259" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.669220] env[62552]: DEBUG oslo_concurrency.lockutils [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.671130] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b17890-33e5-4e3d-903a-0f027ded0b64 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.678820] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1131.679075] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e148f5d-7288-4152-900b-1fde7e6748d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.687991] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1131.687991] env[62552]: value = "task-1240180" [ 1131.687991] env[62552]: _type = "Task" [ 1131.687991] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.695753] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240180, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.175828] env[62552]: INFO nova.compute.manager [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Detaching volume 1900f85e-47dc-4137-825d-646342d03f76 [ 1132.197779] env[62552]: DEBUG oslo_vmware.api [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240180, 'name': PowerOnVM_Task, 'duration_secs': 0.378616} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.198113] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1132.206632] env[62552]: INFO nova.virt.block_device [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Attempting to driver detach volume 1900f85e-47dc-4137-825d-646342d03f76 from mountpoint /dev/sdb [ 1132.206854] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1132.207094] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267534', 'volume_id': '1900f85e-47dc-4137-825d-646342d03f76', 'name': 'volume-1900f85e-47dc-4137-825d-646342d03f76', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60238708-4d45-4d73-bd8d-2cbc25704259', 'attached_at': '', 'detached_at': '', 'volume_id': '1900f85e-47dc-4137-825d-646342d03f76', 'serial': '1900f85e-47dc-4137-825d-646342d03f76'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1132.207906] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b363ad8a-9259-4fc0-946c-b8b20ef40044 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.230161] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d370f33-c7bf-448f-9f53-827737435949 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.238067] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a0e5b9-4c55-4210-89f0-22c3315f77ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.258066] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5711fbd-9eaa-4b0c-a464-69350b0ff702 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.274507] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] The volume has not been displaced from its original location: [datastore2] volume-1900f85e-47dc-4137-825d-646342d03f76/volume-1900f85e-47dc-4137-825d-646342d03f76.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1132.279748] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1132.280052] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78fa4475-be40-4203-9ff4-8592e09824e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.299776] env[62552]: DEBUG oslo_vmware.api [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1132.299776] env[62552]: value = "task-1240182" [ 1132.299776] env[62552]: _type = "Task" [ 1132.299776] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.308546] env[62552]: DEBUG oslo_vmware.api [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240182, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.810672] env[62552]: DEBUG oslo_vmware.api [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240182, 'name': ReconfigVM_Task, 'duration_secs': 0.263232} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.811074] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1132.815665] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ad78512-6e8e-4ebf-8f4c-8fc049ef77bd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.831395] env[62552]: DEBUG oslo_vmware.api [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1132.831395] env[62552]: value = "task-1240183" [ 1132.831395] env[62552]: _type = "Task" [ 1132.831395] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.839339] env[62552]: DEBUG oslo_vmware.api [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240183, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.208143] env[62552]: INFO nova.compute.manager [None req-1eb9604e-7904-421b-a33e-6f9c8c1e7912 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance to original state: 'active' [ 1133.341509] env[62552]: DEBUG oslo_vmware.api [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240183, 'name': ReconfigVM_Task, 'duration_secs': 0.167456} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.341849] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267534', 'volume_id': '1900f85e-47dc-4137-825d-646342d03f76', 'name': 'volume-1900f85e-47dc-4137-825d-646342d03f76', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60238708-4d45-4d73-bd8d-2cbc25704259', 'attached_at': '', 'detached_at': '', 'volume_id': '1900f85e-47dc-4137-825d-646342d03f76', 'serial': '1900f85e-47dc-4137-825d-646342d03f76'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1133.881088] env[62552]: DEBUG nova.objects.instance [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'flavor' on Instance uuid 60238708-4d45-4d73-bd8d-2cbc25704259 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.887937] env[62552]: DEBUG oslo_concurrency.lockutils [None req-76570aac-d452-4254-8513-fcf6083e94b4 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.218s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.910058] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "68cb79de-e08c-4ceb-acca-e45952f67248" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.910348] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "68cb79de-e08c-4ceb-acca-e45952f67248" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.910555] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "68cb79de-e08c-4ceb-acca-e45952f67248-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.910744] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "68cb79de-e08c-4ceb-acca-e45952f67248-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.910916] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "68cb79de-e08c-4ceb-acca-e45952f67248-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.913228] env[62552]: INFO nova.compute.manager [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Terminating instance [ 1135.416728] env[62552]: DEBUG nova.compute.manager [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1135.416971] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1135.417886] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b402bd8e-6835-4262-bb62-5ed4ade17fdc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.426146] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1135.426409] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a14f8881-445f-453f-8e19-6af72c790e57 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.433175] env[62552]: DEBUG oslo_vmware.api [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1135.433175] env[62552]: value = "task-1240187" [ 1135.433175] env[62552]: _type = "Task" [ 1135.433175] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.441587] env[62552]: DEBUG oslo_vmware.api [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240187, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.898850] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "60238708-4d45-4d73-bd8d-2cbc25704259" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.899237] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.899318] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "60238708-4d45-4d73-bd8d-2cbc25704259-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.899507] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.899687] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.901896] env[62552]: INFO nova.compute.manager [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Terminating instance [ 1135.944828] env[62552]: DEBUG oslo_vmware.api [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240187, 'name': PowerOffVM_Task, 'duration_secs': 0.185139} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.945096] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1135.945284] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1135.945514] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9200595a-9b64-4875-9b7e-049f01275e7c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.012267] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1136.012500] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1136.012692] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleting the datastore file [datastore2] 68cb79de-e08c-4ceb-acca-e45952f67248 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1136.012975] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09095bb3-9c87-4568-b9c2-e55edd3c5c1b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.019572] env[62552]: DEBUG oslo_vmware.api [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1136.019572] env[62552]: value = "task-1240189" [ 1136.019572] env[62552]: _type = "Task" [ 1136.019572] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.028254] env[62552]: DEBUG oslo_vmware.api [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240189, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.405822] env[62552]: DEBUG nova.compute.manager [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1136.405822] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1136.406609] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c945ddb-42c0-499f-b3d8-319af796cdd7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.414680] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1136.414962] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cf7dd96-7a34-4dd5-8c87-286d0d24b452 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.421481] env[62552]: DEBUG oslo_vmware.api [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1136.421481] env[62552]: value = "task-1240190" [ 1136.421481] env[62552]: _type = "Task" [ 1136.421481] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.430962] env[62552]: DEBUG oslo_vmware.api [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.531129] env[62552]: DEBUG oslo_vmware.api [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240189, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145936} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.531528] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1136.531822] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1136.532120] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1136.532396] env[62552]: INFO nova.compute.manager [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1136.532736] env[62552]: DEBUG oslo.service.loopingcall [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1136.533032] env[62552]: DEBUG nova.compute.manager [-] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1136.533198] env[62552]: DEBUG nova.network.neutron [-] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1136.932113] env[62552]: DEBUG oslo_vmware.api [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240190, 'name': PowerOffVM_Task, 'duration_secs': 0.163297} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.932566] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1136.932566] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1136.932781] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb231ec0-57d8-4253-8b51-d76166a32354 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.997177] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1136.997177] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Deleting contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1136.997322] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleting the datastore file [datastore1] 60238708-4d45-4d73-bd8d-2cbc25704259 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1136.997544] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6dc3d87b-c635-4a24-904e-5d978f96e72c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.005030] env[62552]: DEBUG oslo_vmware.api [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for the task: (returnval){ [ 1137.005030] env[62552]: value = "task-1240192" [ 1137.005030] env[62552]: _type = "Task" [ 1137.005030] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.012804] env[62552]: DEBUG oslo_vmware.api [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240192, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.044054] env[62552]: DEBUG nova.compute.manager [req-5cc4623b-fd5a-4be7-8525-a2ba9f0bc820 req-711c49ee-46a0-4452-b2d3-7ffc6a8a2f42 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Received event network-vif-deleted-4f6cbf66-fc98-4868-82ef-2f234be029cb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1137.044054] env[62552]: INFO nova.compute.manager [req-5cc4623b-fd5a-4be7-8525-a2ba9f0bc820 req-711c49ee-46a0-4452-b2d3-7ffc6a8a2f42 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Neutron deleted interface 4f6cbf66-fc98-4868-82ef-2f234be029cb; detaching it from the instance and deleting it from the info cache [ 1137.044400] env[62552]: DEBUG nova.network.neutron [req-5cc4623b-fd5a-4be7-8525-a2ba9f0bc820 req-711c49ee-46a0-4452-b2d3-7ffc6a8a2f42 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.069714] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "158d4227-90ca-41ae-821b-efd353928cb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.069714] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.069891] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "158d4227-90ca-41ae-821b-efd353928cb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.070167] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.070407] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.073918] env[62552]: INFO nova.compute.manager [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Terminating instance [ 1137.515574] env[62552]: DEBUG nova.network.neutron [-] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.516767] env[62552]: DEBUG oslo_vmware.api [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Task: {'id': task-1240192, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131592} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.517213] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1137.517405] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Deleted contents of the VM from datastore datastore1 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1137.517588] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1137.517764] env[62552]: INFO nova.compute.manager [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1137.518015] env[62552]: DEBUG oslo.service.loopingcall [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1137.518222] env[62552]: DEBUG nova.compute.manager [-] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1137.518305] env[62552]: DEBUG nova.network.neutron [-] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1137.546886] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41076a9f-e2ea-4456-bb0b-c92148d5bed5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.558708] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d2c71c-e305-46be-9e84-f602a5ddbb36 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.578532] env[62552]: DEBUG nova.compute.manager [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1137.579152] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1137.579620] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d4542d-93e0-47fd-b4da-dda6ab155f35 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.588643] env[62552]: DEBUG nova.compute.manager [req-5cc4623b-fd5a-4be7-8525-a2ba9f0bc820 req-711c49ee-46a0-4452-b2d3-7ffc6a8a2f42 service nova] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Detach interface failed, port_id=4f6cbf66-fc98-4868-82ef-2f234be029cb, reason: Instance 68cb79de-e08c-4ceb-acca-e45952f67248 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1137.594699] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.596937] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0aff8dbd-495d-4b12-9237-0a8e97e32332 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.604556] env[62552]: DEBUG oslo_vmware.api [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1137.604556] env[62552]: value = "task-1240193" [ 1137.604556] env[62552]: _type = "Task" [ 1137.604556] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.614335] env[62552]: DEBUG oslo_vmware.api [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240193, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.017480] env[62552]: INFO nova.compute.manager [-] [instance: 68cb79de-e08c-4ceb-acca-e45952f67248] Took 1.48 seconds to deallocate network for instance. [ 1138.114954] env[62552]: DEBUG oslo_vmware.api [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240193, 'name': PowerOffVM_Task, 'duration_secs': 0.210724} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.115335] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1138.115521] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1138.115787] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-613a617e-0113-4dad-b3fb-1b161951231f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.126068] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.162783] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "a7f32272-20db-4291-8ef8-bce7f56b6893" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.163755] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.193134] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1138.193558] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1138.193653] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleting the datastore file [datastore2] 158d4227-90ca-41ae-821b-efd353928cb2 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.193951] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-68710ab0-2b66-415c-b619-ec065baedb3a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.201473] env[62552]: DEBUG oslo_vmware.api [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for the task: (returnval){ [ 1138.201473] env[62552]: value = "task-1240195" [ 1138.201473] env[62552]: _type = "Task" [ 1138.201473] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.210492] env[62552]: DEBUG oslo_vmware.api [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240195, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.507670] env[62552]: DEBUG nova.network.neutron [-] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.522988] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.523585] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.523961] env[62552]: DEBUG nova.objects.instance [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'resources' on Instance uuid 68cb79de-e08c-4ceb-acca-e45952f67248 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.629732] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Getting list of instances from cluster (obj){ [ 1138.629732] env[62552]: value = "domain-c8" [ 1138.629732] env[62552]: _type = "ClusterComputeResource" [ 1138.629732] env[62552]: } {{(pid=62552) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1138.630473] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c03bcf0-a5c8-4bf7-b52a-b7a00e5feb39 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.647274] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Got total of 2 instances {{(pid=62552) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1138.647448] env[62552]: WARNING nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] While synchronizing instance power states, found 5 instances in the database and 2 instances on the hypervisor. [ 1138.647590] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Triggering sync for uuid 6a5b5f4e-0c84-447d-a3da-8258512abb7c {{(pid=62552) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1138.647847] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Triggering sync for uuid 158d4227-90ca-41ae-821b-efd353928cb2 {{(pid=62552) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1138.648035] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Triggering sync for uuid 68cb79de-e08c-4ceb-acca-e45952f67248 {{(pid=62552) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1138.648191] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Triggering sync for uuid 60238708-4d45-4d73-bd8d-2cbc25704259 {{(pid=62552) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1138.648335] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Triggering sync for uuid 888ed8fb-18e6-4f69-a27c-bd24c19e0a12 {{(pid=62552) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1138.648669] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.648901] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.649548] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "158d4227-90ca-41ae-821b-efd353928cb2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.649548] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "68cb79de-e08c-4ceb-acca-e45952f67248" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.649694] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "60238708-4d45-4d73-bd8d-2cbc25704259" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.649783] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.649961] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.650826] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00223d90-0eb7-4e78-a295-d3ec5bced778 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.653861] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d97f6d9-205d-4203-90af-b3296c7c33f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.667270] env[62552]: DEBUG nova.compute.manager [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1138.713727] env[62552]: DEBUG oslo_vmware.api [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Task: {'id': task-1240195, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136069} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.713997] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.714208] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1138.714390] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1138.714568] env[62552]: INFO nova.compute.manager [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1138.714809] env[62552]: DEBUG oslo.service.loopingcall [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.715012] env[62552]: DEBUG nova.compute.manager [-] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1138.715113] env[62552]: DEBUG nova.network.neutron [-] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1139.010439] env[62552]: INFO nova.compute.manager [-] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Took 1.49 seconds to deallocate network for instance. [ 1139.076047] env[62552]: DEBUG nova.compute.manager [req-ff0819bd-d9f0-4ce6-ab53-b7ce5521b13c req-95e0aaff-7cae-4eda-9458-cb43a0114e5e service nova] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] Received event network-vif-deleted-a1d4afdd-37b2-4b59-b447-0ca443e4c7f6 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1139.137164] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9079f65b-945b-402a-88cc-8585a02de218 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.145965] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7745877-37c3-4b67-a48f-bd6cb29812ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.186125] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.536s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.189150] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.539s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.192532] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf545573-549d-4ef5-95c9-13cc461f06a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.201977] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2b9dbd-50e6-4694-a3a0-23b9645b0b1e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.210271] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.221630] env[62552]: DEBUG nova.compute.provider_tree [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.517096] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.724032] env[62552]: DEBUG nova.scheduler.client.report [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1139.974359] env[62552]: DEBUG nova.network.neutron [-] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.229108] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.705s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.231443] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.021s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.233107] env[62552]: INFO nova.compute.claims [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1140.249258] env[62552]: INFO nova.scheduler.client.report [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted allocations for instance 68cb79de-e08c-4ceb-acca-e45952f67248 [ 1140.477235] env[62552]: INFO nova.compute.manager [-] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Took 1.76 seconds to deallocate network for instance. [ 1140.756424] env[62552]: DEBUG oslo_concurrency.lockutils [None req-50ca7637-9171-470a-ad77-b383297a844b tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "68cb79de-e08c-4ceb-acca-e45952f67248" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.846s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.757409] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "68cb79de-e08c-4ceb-acca-e45952f67248" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.108s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.757634] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-366e34c6-a98c-403c-918f-1aa0a9bf7c85 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.768192] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2da46b-e449-4b8a-bf5a-b9046b532299 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.984347] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.100399] env[62552]: DEBUG nova.compute.manager [req-df203f09-33f4-4334-abc2-2fa972f1fc79 req-c0a6f193-d4f6-45e5-a778-795bd66936df service nova] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] Received event network-vif-deleted-dd1028f0-d5a9-48be-84e3-9401cde0a1ff {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1141.296151] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "68cb79de-e08c-4ceb-acca-e45952f67248" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.539s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.306810] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb78db7-a94d-4b10-8630-6d024f230a20 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.314474] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77aca75-1e12-43fb-bef7-9dd9a2f7444a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.345158] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76f2217-d2ec-4560-8612-516e17ced30e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.352049] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f77c87c-c465-4896-8602-9d9142aa2042 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.364997] env[62552]: DEBUG nova.compute.provider_tree [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.867936] env[62552]: DEBUG nova.scheduler.client.report [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1142.027366] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "44b075d0-66ac-4b49-928d-d15db17a282e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.027616] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "44b075d0-66ac-4b49-928d-d15db17a282e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.373090] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.141s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.373462] env[62552]: DEBUG nova.compute.manager [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1142.376541] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.860s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.376784] env[62552]: DEBUG nova.objects.instance [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lazy-loading 'resources' on Instance uuid 60238708-4d45-4d73-bd8d-2cbc25704259 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.529565] env[62552]: DEBUG nova.compute.manager [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1142.879872] env[62552]: DEBUG nova.compute.utils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1142.884040] env[62552]: DEBUG nova.compute.manager [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1142.884155] env[62552]: DEBUG nova.network.neutron [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1142.930840] env[62552]: DEBUG nova.policy [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb432303c2394f8795672ceaf02f28aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a0bc011d6794602b2bbe1fc01e4c8b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1142.965649] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fffe1e29-f20a-4aab-83de-517171f60179 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.973475] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2406f84-ed97-4112-b2c4-283b875961e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.003849] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d85fe06-f6ef-486c-a3bb-903d80d637b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.010484] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4493278-2634-41cc-b7f7-34a921c72ec7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.023340] env[62552]: DEBUG nova.compute.provider_tree [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.048521] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.212951] env[62552]: DEBUG nova.network.neutron [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Successfully created port: 1961d456-c180-4993-a8b5-33fab8b7e18c {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1143.384770] env[62552]: DEBUG nova.compute.manager [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1143.527148] env[62552]: DEBUG nova.scheduler.client.report [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1143.890232] env[62552]: INFO nova.virt.block_device [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Booting with volume b48a8f1d-dc4a-4b43-ace9-214a64d673da at /dev/sda [ 1143.925520] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a5cf2fc-34ad-4cfd-bf87-df1f1887f5c5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.936202] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167c57fc-9882-4bde-a773-05f0c9bdedd7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.960738] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a83de105-a025-4c1b-aa03-0c1fb918be6c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.968511] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb314da9-c04e-42d4-9fb4-a0c7b60a3aa0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.992599] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f131314-8823-44d7-89b7-daf12b380ed6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.998490] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a87da5b-0ac0-4cd4-8a12-b6fbb9b8cd18 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.011527] env[62552]: DEBUG nova.virt.block_device [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating existing volume attachment record: 220ba93a-d856-4081-8786-1ba8d7aff6c5 {{(pid=62552) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1144.033319] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.657s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.035476] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.051s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.035710] env[62552]: DEBUG nova.objects.instance [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lazy-loading 'resources' on Instance uuid 158d4227-90ca-41ae-821b-efd353928cb2 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.054929] env[62552]: INFO nova.scheduler.client.report [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Deleted allocations for instance 60238708-4d45-4d73-bd8d-2cbc25704259 [ 1144.564454] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c43e20ae-0247-4683-a802-070e7a7ee8a2 tempest-AttachVolumeNegativeTest-1459990734 tempest-AttachVolumeNegativeTest-1459990734-project-member] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.665s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.565451] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.916s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.565649] env[62552]: INFO nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 60238708-4d45-4d73-bd8d-2cbc25704259] During sync_power_state the instance has a pending task (deleting). Skip. [ 1144.565825] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "60238708-4d45-4d73-bd8d-2cbc25704259" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.624174] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b872ae9b-460e-419c-90a9-285530e9fecf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.632464] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5237fd5-13f4-4289-98ce-2bda953f0bdf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.663479] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af70452c-00c5-442c-87ac-227a2510fb94 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.666776] env[62552]: DEBUG nova.compute.manager [req-d4c02896-40c8-4002-a084-cf66d9542ed7 req-47468217-a7e6-41cd-8e55-9df0cf11ce28 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Received event network-vif-plugged-1961d456-c180-4993-a8b5-33fab8b7e18c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1144.666988] env[62552]: DEBUG oslo_concurrency.lockutils [req-d4c02896-40c8-4002-a084-cf66d9542ed7 req-47468217-a7e6-41cd-8e55-9df0cf11ce28 service nova] Acquiring lock "a7f32272-20db-4291-8ef8-bce7f56b6893-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.667210] env[62552]: DEBUG oslo_concurrency.lockutils [req-d4c02896-40c8-4002-a084-cf66d9542ed7 req-47468217-a7e6-41cd-8e55-9df0cf11ce28 service nova] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.667381] env[62552]: DEBUG oslo_concurrency.lockutils [req-d4c02896-40c8-4002-a084-cf66d9542ed7 req-47468217-a7e6-41cd-8e55-9df0cf11ce28 service nova] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.668029] env[62552]: DEBUG nova.compute.manager [req-d4c02896-40c8-4002-a084-cf66d9542ed7 req-47468217-a7e6-41cd-8e55-9df0cf11ce28 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] No waiting events found dispatching network-vif-plugged-1961d456-c180-4993-a8b5-33fab8b7e18c {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1144.668029] env[62552]: WARNING nova.compute.manager [req-d4c02896-40c8-4002-a084-cf66d9542ed7 req-47468217-a7e6-41cd-8e55-9df0cf11ce28 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Received unexpected event network-vif-plugged-1961d456-c180-4993-a8b5-33fab8b7e18c for instance with vm_state building and task_state block_device_mapping. [ 1144.673349] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddfde79-4fac-4f23-b0ba-2688b3ceb522 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.688250] env[62552]: DEBUG nova.compute.provider_tree [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1144.753052] env[62552]: DEBUG nova.network.neutron [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Successfully updated port: 1961d456-c180-4993-a8b5-33fab8b7e18c {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1145.191261] env[62552]: DEBUG nova.scheduler.client.report [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1145.262907] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.263076] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.263228] env[62552]: DEBUG nova.network.neutron [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1145.696089] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.660s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.697916] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.649s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.699469] env[62552]: INFO nova.compute.claims [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1145.715790] env[62552]: INFO nova.scheduler.client.report [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Deleted allocations for instance 158d4227-90ca-41ae-821b-efd353928cb2 [ 1145.816110] env[62552]: DEBUG nova.network.neutron [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1145.947673] env[62552]: DEBUG nova.network.neutron [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance_info_cache with network_info: [{"id": "1961d456-c180-4993-a8b5-33fab8b7e18c", "address": "fa:16:3e:ce:0d:1a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1961d456-c1", "ovs_interfaceid": "1961d456-c180-4993-a8b5-33fab8b7e18c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.096639] env[62552]: DEBUG nova.compute.manager [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1146.097190] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1146.097414] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1146.097637] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1146.097971] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1146.098087] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1146.098179] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1146.098385] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1146.098550] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1146.098724] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1146.098896] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1146.099088] env[62552]: DEBUG nova.virt.hardware [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.099929] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56782517-988f-40e5-b629-0f2c253b831c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.108128] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa054cf2-e4cf-4808-906d-11d2e7b85301 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.228136] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d1835706-eb8b-4cbd-b0e6-06ba8d13ebaf tempest-AttachVolumeShelveTestJSON-1589972985 tempest-AttachVolumeShelveTestJSON-1589972985-project-member] Lock "158d4227-90ca-41ae-821b-efd353928cb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.158s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.230272] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "158d4227-90ca-41ae-821b-efd353928cb2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.581s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.230537] env[62552]: INFO nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 158d4227-90ca-41ae-821b-efd353928cb2] During sync_power_state the instance has a pending task (deleting). Skip. [ 1146.230838] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "158d4227-90ca-41ae-821b-efd353928cb2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.450512] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.450857] env[62552]: DEBUG nova.compute.manager [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Instance network_info: |[{"id": "1961d456-c180-4993-a8b5-33fab8b7e18c", "address": "fa:16:3e:ce:0d:1a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1961d456-c1", "ovs_interfaceid": "1961d456-c180-4993-a8b5-33fab8b7e18c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1146.451323] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:0d:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1961d456-c180-4993-a8b5-33fab8b7e18c', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1146.458792] env[62552]: DEBUG oslo.service.loopingcall [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1146.459033] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1146.459278] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0bf9563d-621c-45a8-b1b7-33d09b86c3f0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.479154] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1146.479154] env[62552]: value = "task-1240197" [ 1146.479154] env[62552]: _type = "Task" [ 1146.479154] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.486358] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240197, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.683480] env[62552]: DEBUG nova.compute.manager [req-d0a850fd-f025-4f19-8c8e-0845f400822c req-95d55d80-cb43-4106-a192-769c43b3589b service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Received event network-changed-1961d456-c180-4993-a8b5-33fab8b7e18c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1146.683626] env[62552]: DEBUG nova.compute.manager [req-d0a850fd-f025-4f19-8c8e-0845f400822c req-95d55d80-cb43-4106-a192-769c43b3589b service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Refreshing instance network info cache due to event network-changed-1961d456-c180-4993-a8b5-33fab8b7e18c. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1146.683900] env[62552]: DEBUG oslo_concurrency.lockutils [req-d0a850fd-f025-4f19-8c8e-0845f400822c req-95d55d80-cb43-4106-a192-769c43b3589b service nova] Acquiring lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.684076] env[62552]: DEBUG oslo_concurrency.lockutils [req-d0a850fd-f025-4f19-8c8e-0845f400822c req-95d55d80-cb43-4106-a192-769c43b3589b service nova] Acquired lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.684250] env[62552]: DEBUG nova.network.neutron [req-d0a850fd-f025-4f19-8c8e-0845f400822c req-95d55d80-cb43-4106-a192-769c43b3589b service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Refreshing network info cache for port 1961d456-c180-4993-a8b5-33fab8b7e18c {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1146.782739] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3763af4f-752d-4b51-bdf8-1261705f2073 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.790690] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d3dd99-d0ae-400c-8796-eeb8272df964 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.822284] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977260f2-c497-4218-95bc-28e473fc4683 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.830042] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a86fe69-cbb9-4271-a5fa-6ce499e4463c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.847533] env[62552]: DEBUG nova.compute.provider_tree [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1146.989134] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240197, 'name': CreateVM_Task, 'duration_secs': 0.28864} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.989307] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1146.989997] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267535', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'name': 'volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a7f32272-20db-4291-8ef8-bce7f56b6893', 'attached_at': '', 'detached_at': '', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'serial': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da'}, 'disk_bus': None, 'boot_index': 0, 'attachment_id': '220ba93a-d856-4081-8786-1ba8d7aff6c5', 'volume_type': None}], 'swap': None} {{(pid=62552) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1146.990238] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Root volume attach. Driver type: vmdk {{(pid=62552) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1146.990979] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9f7da9-6f1b-4f24-9dc3-9a7f33d05276 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.997911] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a294b2-5027-47a1-9610-d9ad5b26072d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.003831] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24884ff0-d536-4db7-92ee-9c34e26c3a82 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.009327] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-7b61d325-7aa9-4712-b6a4-f96fb6181dd6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.015573] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1147.015573] env[62552]: value = "task-1240198" [ 1147.015573] env[62552]: _type = "Task" [ 1147.015573] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.023500] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240198, 'name': RelocateVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.350925] env[62552]: DEBUG nova.scheduler.client.report [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1147.398973] env[62552]: DEBUG nova.network.neutron [req-d0a850fd-f025-4f19-8c8e-0845f400822c req-95d55d80-cb43-4106-a192-769c43b3589b service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updated VIF entry in instance network info cache for port 1961d456-c180-4993-a8b5-33fab8b7e18c. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1147.399366] env[62552]: DEBUG nova.network.neutron [req-d0a850fd-f025-4f19-8c8e-0845f400822c req-95d55d80-cb43-4106-a192-769c43b3589b service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance_info_cache with network_info: [{"id": "1961d456-c180-4993-a8b5-33fab8b7e18c", "address": "fa:16:3e:ce:0d:1a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1961d456-c1", "ovs_interfaceid": "1961d456-c180-4993-a8b5-33fab8b7e18c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.525265] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240198, 'name': RelocateVM_Task, 'duration_secs': 0.355163} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.525503] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Volume attach. Driver type: vmdk {{(pid=62552) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1147.525710] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267535', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'name': 'volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a7f32272-20db-4291-8ef8-bce7f56b6893', 'attached_at': '', 'detached_at': '', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'serial': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1147.526502] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e92450-7c9c-4528-a700-c4d52b906a26 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.541691] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4332423-620d-4c5d-95d7-9e7741ede380 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.562390] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da/volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1147.562631] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-415e010d-c329-48da-a03f-d8d1299615ff {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.582465] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1147.582465] env[62552]: value = "task-1240199" [ 1147.582465] env[62552]: _type = "Task" [ 1147.582465] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.589405] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240199, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.856095] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.158s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.856698] env[62552]: DEBUG nova.compute.manager [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1147.902028] env[62552]: DEBUG oslo_concurrency.lockutils [req-d0a850fd-f025-4f19-8c8e-0845f400822c req-95d55d80-cb43-4106-a192-769c43b3589b service nova] Releasing lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.092130] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240199, 'name': ReconfigVM_Task, 'duration_secs': 0.247999} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.092460] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da/volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.097140] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fdbef0d0-83b0-4327-bc7b-f0280abe2a26 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.123315] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1148.123315] env[62552]: value = "task-1240202" [ 1148.123315] env[62552]: _type = "Task" [ 1148.123315] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.132441] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240202, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.362549] env[62552]: DEBUG nova.compute.utils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1148.363786] env[62552]: DEBUG nova.compute.manager [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1148.363970] env[62552]: DEBUG nova.network.neutron [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1148.402474] env[62552]: DEBUG nova.policy [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bdea5d6f50634b5e89852744db56db67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8495bac65967441996ecca1fd22a4da4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1148.633641] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240202, 'name': ReconfigVM_Task, 'duration_secs': 0.114765} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.634035] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267535', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'name': 'volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a7f32272-20db-4291-8ef8-bce7f56b6893', 'attached_at': '', 'detached_at': '', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'serial': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da'} {{(pid=62552) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1148.634578] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e358ef6-6f7b-4240-8961-660a63493d84 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.641185] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1148.641185] env[62552]: value = "task-1240203" [ 1148.641185] env[62552]: _type = "Task" [ 1148.641185] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.649526] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240203, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.665242] env[62552]: DEBUG nova.network.neutron [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Successfully created port: dd8cb7f7-fc67-4477-bdd4-27617d60e42e {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1148.867374] env[62552]: DEBUG nova.compute.manager [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1149.151051] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240203, 'name': Rename_Task, 'duration_secs': 0.12214} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.151353] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1149.151612] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-929e130c-cf6d-45d9-8e87-85a6613477bb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.157868] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1149.157868] env[62552]: value = "task-1240204" [ 1149.157868] env[62552]: _type = "Task" [ 1149.157868] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.165186] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240204, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.668281] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240204, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.879193] env[62552]: DEBUG nova.compute.manager [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1149.907203] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1149.907480] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1149.907642] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1149.907826] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1149.907976] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1149.908149] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1149.908362] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1149.908523] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1149.908695] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1149.908862] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1149.909051] env[62552]: DEBUG nova.virt.hardware [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1149.909918] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5febe74-82b4-45f0-b276-ba8cbc293acd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.917758] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc11911a-5ec1-4fa3-8f24-dc62b49dfd6a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.064905] env[62552]: DEBUG nova.compute.manager [req-ac550347-4fb9-4fe6-946c-d21bea86c6eb req-a1dd107d-dee9-439e-b4f2-5acec0a0a049 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Received event network-vif-plugged-dd8cb7f7-fc67-4477-bdd4-27617d60e42e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1150.065162] env[62552]: DEBUG oslo_concurrency.lockutils [req-ac550347-4fb9-4fe6-946c-d21bea86c6eb req-a1dd107d-dee9-439e-b4f2-5acec0a0a049 service nova] Acquiring lock "44b075d0-66ac-4b49-928d-d15db17a282e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.065379] env[62552]: DEBUG oslo_concurrency.lockutils [req-ac550347-4fb9-4fe6-946c-d21bea86c6eb req-a1dd107d-dee9-439e-b4f2-5acec0a0a049 service nova] Lock "44b075d0-66ac-4b49-928d-d15db17a282e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.065567] env[62552]: DEBUG oslo_concurrency.lockutils [req-ac550347-4fb9-4fe6-946c-d21bea86c6eb req-a1dd107d-dee9-439e-b4f2-5acec0a0a049 service nova] Lock "44b075d0-66ac-4b49-928d-d15db17a282e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.065742] env[62552]: DEBUG nova.compute.manager [req-ac550347-4fb9-4fe6-946c-d21bea86c6eb req-a1dd107d-dee9-439e-b4f2-5acec0a0a049 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] No waiting events found dispatching network-vif-plugged-dd8cb7f7-fc67-4477-bdd4-27617d60e42e {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1150.065923] env[62552]: WARNING nova.compute.manager [req-ac550347-4fb9-4fe6-946c-d21bea86c6eb req-a1dd107d-dee9-439e-b4f2-5acec0a0a049 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Received unexpected event network-vif-plugged-dd8cb7f7-fc67-4477-bdd4-27617d60e42e for instance with vm_state building and task_state spawning. [ 1150.147648] env[62552]: DEBUG nova.network.neutron [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Successfully updated port: dd8cb7f7-fc67-4477-bdd4-27617d60e42e {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1150.169836] env[62552]: DEBUG oslo_vmware.api [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240204, 'name': PowerOnVM_Task, 'duration_secs': 0.646734} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.169836] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1150.169836] env[62552]: INFO nova.compute.manager [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Took 4.07 seconds to spawn the instance on the hypervisor. [ 1150.169836] env[62552]: DEBUG nova.compute.manager [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1150.170624] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1cfe51-80cf-4478-8efa-d3daabaa875f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.652117] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.652334] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.652429] env[62552]: DEBUG nova.network.neutron [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1150.690311] env[62552]: INFO nova.compute.manager [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Took 11.50 seconds to build instance. [ 1151.187041] env[62552]: DEBUG nova.network.neutron [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1151.192777] env[62552]: DEBUG oslo_concurrency.lockutils [None req-0536049f-c401-4463-811c-8b46d5469cf8 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.030s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.395142] env[62552]: DEBUG nova.network.neutron [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Updating instance_info_cache with network_info: [{"id": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "address": "fa:16:3e:75:d0:12", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd8cb7f7-fc", "ovs_interfaceid": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.702940] env[62552]: DEBUG nova.compute.manager [req-0320f1a6-cfc0-4f73-8b9c-58147dedc669 req-a4bdacf4-2600-43ee-83c7-baa64573c1e4 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Received event network-changed-5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1151.703178] env[62552]: DEBUG nova.compute.manager [req-0320f1a6-cfc0-4f73-8b9c-58147dedc669 req-a4bdacf4-2600-43ee-83c7-baa64573c1e4 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Refreshing instance network info cache due to event network-changed-5e69a23a-a63b-4d24-9729-859b537f945b. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1151.703400] env[62552]: DEBUG oslo_concurrency.lockutils [req-0320f1a6-cfc0-4f73-8b9c-58147dedc669 req-a4bdacf4-2600-43ee-83c7-baa64573c1e4 service nova] Acquiring lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.703548] env[62552]: DEBUG oslo_concurrency.lockutils [req-0320f1a6-cfc0-4f73-8b9c-58147dedc669 req-a4bdacf4-2600-43ee-83c7-baa64573c1e4 service nova] Acquired lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.703715] env[62552]: DEBUG nova.network.neutron [req-0320f1a6-cfc0-4f73-8b9c-58147dedc669 req-a4bdacf4-2600-43ee-83c7-baa64573c1e4 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Refreshing network info cache for port 5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1151.903014] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.903014] env[62552]: DEBUG nova.compute.manager [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Instance network_info: |[{"id": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "address": "fa:16:3e:75:d0:12", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd8cb7f7-fc", "ovs_interfaceid": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1151.903014] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:d0:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd8cb7f7-fc67-4477-bdd4-27617d60e42e', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1151.911599] env[62552]: DEBUG oslo.service.loopingcall [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1151.911599] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1151.911971] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f86d0e4-95b3-4323-95d0-1534b82378bf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.933259] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1151.933259] env[62552]: value = "task-1240206" [ 1151.933259] env[62552]: _type = "Task" [ 1151.933259] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.941837] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240206, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.104144] env[62552]: DEBUG nova.compute.manager [req-d5ca6d24-7c9b-43ab-8620-428849114daf req-e8371040-6c81-4c8f-92ac-1fa6850c4465 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Received event network-changed-dd8cb7f7-fc67-4477-bdd4-27617d60e42e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1152.104353] env[62552]: DEBUG nova.compute.manager [req-d5ca6d24-7c9b-43ab-8620-428849114daf req-e8371040-6c81-4c8f-92ac-1fa6850c4465 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Refreshing instance network info cache due to event network-changed-dd8cb7f7-fc67-4477-bdd4-27617d60e42e. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1152.104714] env[62552]: DEBUG oslo_concurrency.lockutils [req-d5ca6d24-7c9b-43ab-8620-428849114daf req-e8371040-6c81-4c8f-92ac-1fa6850c4465 service nova] Acquiring lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.104769] env[62552]: DEBUG oslo_concurrency.lockutils [req-d5ca6d24-7c9b-43ab-8620-428849114daf req-e8371040-6c81-4c8f-92ac-1fa6850c4465 service nova] Acquired lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.104898] env[62552]: DEBUG nova.network.neutron [req-d5ca6d24-7c9b-43ab-8620-428849114daf req-e8371040-6c81-4c8f-92ac-1fa6850c4465 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Refreshing network info cache for port dd8cb7f7-fc67-4477-bdd4-27617d60e42e {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1152.443451] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240206, 'name': CreateVM_Task, 'duration_secs': 0.328792} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.443748] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1152.444448] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.444666] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.445050] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1152.445349] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27f72897-3f38-4301-babb-e56718d84fec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.452822] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1152.452822] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b3dcbf-720e-9d65-b48d-9583c08184e8" [ 1152.452822] env[62552]: _type = "Task" [ 1152.452822] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.461542] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b3dcbf-720e-9d65-b48d-9583c08184e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.487613] env[62552]: DEBUG nova.network.neutron [req-0320f1a6-cfc0-4f73-8b9c-58147dedc669 req-a4bdacf4-2600-43ee-83c7-baa64573c1e4 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updated VIF entry in instance network info cache for port 5e69a23a-a63b-4d24-9729-859b537f945b. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1152.487992] env[62552]: DEBUG nova.network.neutron [req-0320f1a6-cfc0-4f73-8b9c-58147dedc669 req-a4bdacf4-2600-43ee-83c7-baa64573c1e4 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updating instance_info_cache with network_info: [{"id": "5e69a23a-a63b-4d24-9729-859b537f945b", "address": "fa:16:3e:0b:c1:98", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e69a23a-a6", "ovs_interfaceid": "5e69a23a-a63b-4d24-9729-859b537f945b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.829883] env[62552]: DEBUG nova.network.neutron [req-d5ca6d24-7c9b-43ab-8620-428849114daf req-e8371040-6c81-4c8f-92ac-1fa6850c4465 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Updated VIF entry in instance network info cache for port dd8cb7f7-fc67-4477-bdd4-27617d60e42e. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1152.830286] env[62552]: DEBUG nova.network.neutron [req-d5ca6d24-7c9b-43ab-8620-428849114daf req-e8371040-6c81-4c8f-92ac-1fa6850c4465 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Updating instance_info_cache with network_info: [{"id": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "address": "fa:16:3e:75:d0:12", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd8cb7f7-fc", "ovs_interfaceid": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.968101] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b3dcbf-720e-9d65-b48d-9583c08184e8, 'name': SearchDatastore_Task, 'duration_secs': 0.021067} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.968479] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.968724] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1152.968966] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.969178] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.969393] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1152.970010] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7a98598-ef00-4892-bd39-5e022c6c6242 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.979599] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1152.979815] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1152.980578] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fce2dad2-7c3a-4f1e-a031-36010e538a2e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.986056] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1152.986056] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5230e2aa-5b2a-9b32-88fa-32ea61e1d9a4" [ 1152.986056] env[62552]: _type = "Task" [ 1152.986056] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.991037] env[62552]: DEBUG oslo_concurrency.lockutils [req-0320f1a6-cfc0-4f73-8b9c-58147dedc669 req-a4bdacf4-2600-43ee-83c7-baa64573c1e4 service nova] Releasing lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.994282] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5230e2aa-5b2a-9b32-88fa-32ea61e1d9a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.183216] env[62552]: DEBUG nova.compute.manager [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Stashing vm_state: active {{(pid=62552) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 1153.333187] env[62552]: DEBUG oslo_concurrency.lockutils [req-d5ca6d24-7c9b-43ab-8620-428849114daf req-e8371040-6c81-4c8f-92ac-1fa6850c4465 service nova] Releasing lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.497313] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5230e2aa-5b2a-9b32-88fa-32ea61e1d9a4, 'name': SearchDatastore_Task, 'duration_secs': 0.00922} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.498128] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3b526dd-8930-4b4d-a4f6-43cb679b37d3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.502961] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1153.502961] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521280b7-98e7-224a-0633-eb085b14426b" [ 1153.502961] env[62552]: _type = "Task" [ 1153.502961] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.510736] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521280b7-98e7-224a-0633-eb085b14426b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.704573] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.705221] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.733851] env[62552]: DEBUG nova.compute.manager [req-48c586f8-d4c5-41ec-a20d-eca543833599 req-a15d2138-a919-4c85-ac6b-c1c5b9119382 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Received event network-changed-1961d456-c180-4993-a8b5-33fab8b7e18c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1153.733851] env[62552]: DEBUG nova.compute.manager [req-48c586f8-d4c5-41ec-a20d-eca543833599 req-a15d2138-a919-4c85-ac6b-c1c5b9119382 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Refreshing instance network info cache due to event network-changed-1961d456-c180-4993-a8b5-33fab8b7e18c. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1153.733851] env[62552]: DEBUG oslo_concurrency.lockutils [req-48c586f8-d4c5-41ec-a20d-eca543833599 req-a15d2138-a919-4c85-ac6b-c1c5b9119382 service nova] Acquiring lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.733851] env[62552]: DEBUG oslo_concurrency.lockutils [req-48c586f8-d4c5-41ec-a20d-eca543833599 req-a15d2138-a919-4c85-ac6b-c1c5b9119382 service nova] Acquired lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.733851] env[62552]: DEBUG nova.network.neutron [req-48c586f8-d4c5-41ec-a20d-eca543833599 req-a15d2138-a919-4c85-ac6b-c1c5b9119382 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Refreshing network info cache for port 1961d456-c180-4993-a8b5-33fab8b7e18c {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1154.015644] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]521280b7-98e7-224a-0633-eb085b14426b, 'name': SearchDatastore_Task, 'duration_secs': 0.04256} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.015644] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.015644] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 44b075d0-66ac-4b49-928d-d15db17a282e/44b075d0-66ac-4b49-928d-d15db17a282e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1154.015644] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5fde7245-5e2e-4490-9444-be7ab80ecf44 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.022443] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1154.022443] env[62552]: value = "task-1240207" [ 1154.022443] env[62552]: _type = "Task" [ 1154.022443] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.030190] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240207, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.211227] env[62552]: INFO nova.compute.claims [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1154.534609] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240207, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.693731] env[62552]: DEBUG nova.network.neutron [req-48c586f8-d4c5-41ec-a20d-eca543833599 req-a15d2138-a919-4c85-ac6b-c1c5b9119382 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updated VIF entry in instance network info cache for port 1961d456-c180-4993-a8b5-33fab8b7e18c. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1154.694217] env[62552]: DEBUG nova.network.neutron [req-48c586f8-d4c5-41ec-a20d-eca543833599 req-a15d2138-a919-4c85-ac6b-c1c5b9119382 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance_info_cache with network_info: [{"id": "1961d456-c180-4993-a8b5-33fab8b7e18c", "address": "fa:16:3e:ce:0d:1a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1961d456-c1", "ovs_interfaceid": "1961d456-c180-4993-a8b5-33fab8b7e18c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.718051] env[62552]: INFO nova.compute.resource_tracker [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating resource usage from migration 02e864b8-cba0-41c6-b216-564df74dfd9e [ 1154.792229] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e22579b-c5ae-4d44-8e9f-62e99454b720 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.799964] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57280792-31cd-4ccd-9b46-6ab634b8975b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.834145] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c837f0b7-de01-4f0e-b7d1-c9a874b253c7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.841815] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d95862-4be4-43b3-a4fd-6269e06dcca4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.855212] env[62552]: DEBUG nova.compute.provider_tree [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.033979] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240207, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.197492] env[62552]: DEBUG oslo_concurrency.lockutils [req-48c586f8-d4c5-41ec-a20d-eca543833599 req-a15d2138-a919-4c85-ac6b-c1c5b9119382 service nova] Releasing lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.357987] env[62552]: DEBUG nova.scheduler.client.report [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1155.533718] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240207, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.863196] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.158s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.863562] env[62552]: INFO nova.compute.manager [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Migrating [ 1156.036853] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240207, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.655326} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.037118] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 44b075d0-66ac-4b49-928d-d15db17a282e/44b075d0-66ac-4b49-928d-d15db17a282e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1156.037377] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1156.037625] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fab59118-c407-4d78-8222-ba0cb26638d0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.043656] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1156.043656] env[62552]: value = "task-1240208" [ 1156.043656] env[62552]: _type = "Task" [ 1156.043656] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.050832] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240208, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.379131] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.379393] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.379665] env[62552]: DEBUG nova.network.neutron [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1156.553892] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240208, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077608} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.554200] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1156.554962] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2356e718-d11a-4a91-a466-fb3e1ecf2794 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.576456] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 44b075d0-66ac-4b49-928d-d15db17a282e/44b075d0-66ac-4b49-928d-d15db17a282e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.576717] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f63fa2e7-268b-4f2c-b7bc-f4365651c6d1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.598553] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1156.598553] env[62552]: value = "task-1240209" [ 1156.598553] env[62552]: _type = "Task" [ 1156.598553] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.606527] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240209, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.086301] env[62552]: DEBUG nova.network.neutron [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance_info_cache with network_info: [{"id": "1961d456-c180-4993-a8b5-33fab8b7e18c", "address": "fa:16:3e:ce:0d:1a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1961d456-c1", "ovs_interfaceid": "1961d456-c180-4993-a8b5-33fab8b7e18c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.108079] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240209, 'name': ReconfigVM_Task, 'duration_secs': 0.268182} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.108433] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 44b075d0-66ac-4b49-928d-d15db17a282e/44b075d0-66ac-4b49-928d-d15db17a282e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1157.109149] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-639c6692-fe2e-4a37-bd38-b77730db90f4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.116664] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1157.116664] env[62552]: value = "task-1240210" [ 1157.116664] env[62552]: _type = "Task" [ 1157.116664] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.124480] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240210, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.222297] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1157.588878] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.626425] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240210, 'name': Rename_Task, 'duration_secs': 0.139366} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.626696] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1157.626940] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3626ee91-d3c3-4d89-9eef-942c884aa11b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.633390] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1157.633390] env[62552]: value = "task-1240211" [ 1157.633390] env[62552]: _type = "Task" [ 1157.633390] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.640738] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240211, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.697941] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.143663] env[62552]: DEBUG oslo_vmware.api [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240211, 'name': PowerOnVM_Task, 'duration_secs': 0.441987} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.143951] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1158.144180] env[62552]: INFO nova.compute.manager [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Took 8.26 seconds to spawn the instance on the hypervisor. [ 1158.144383] env[62552]: DEBUG nova.compute.manager [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1158.145191] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e81dffa-41e3-4679-b2c0-6e1bdcd8f250 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.660261] env[62552]: INFO nova.compute.manager [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Took 15.63 seconds to build instance. [ 1158.693018] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.697606] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.697757] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1159.104552] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2371d8-51fe-47e1-a975-8c483fecd3e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.122576] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance 'a7f32272-20db-4291-8ef8-bce7f56b6893' progress to 0 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1159.162584] env[62552]: DEBUG oslo_concurrency.lockutils [None req-72142721-157e-4327-a579-008638eeedce tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "44b075d0-66ac-4b49-928d-d15db17a282e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.135s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.628346] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1159.628665] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fefbc73d-dc05-4307-9b40-b233d9c0cf4b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.635515] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1159.635515] env[62552]: value = "task-1240212" [ 1159.635515] env[62552]: _type = "Task" [ 1159.635515] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.643547] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240212, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.716170] env[62552]: DEBUG nova.compute.manager [req-2f253683-d1a4-48cd-ab57-44cfd7411506 req-9d291b01-a4ee-4c8b-8d83-8a68dab5b91f service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Received event network-changed-dd8cb7f7-fc67-4477-bdd4-27617d60e42e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1159.716302] env[62552]: DEBUG nova.compute.manager [req-2f253683-d1a4-48cd-ab57-44cfd7411506 req-9d291b01-a4ee-4c8b-8d83-8a68dab5b91f service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Refreshing instance network info cache due to event network-changed-dd8cb7f7-fc67-4477-bdd4-27617d60e42e. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1159.716513] env[62552]: DEBUG oslo_concurrency.lockutils [req-2f253683-d1a4-48cd-ab57-44cfd7411506 req-9d291b01-a4ee-4c8b-8d83-8a68dab5b91f service nova] Acquiring lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1159.716662] env[62552]: DEBUG oslo_concurrency.lockutils [req-2f253683-d1a4-48cd-ab57-44cfd7411506 req-9d291b01-a4ee-4c8b-8d83-8a68dab5b91f service nova] Acquired lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.716826] env[62552]: DEBUG nova.network.neutron [req-2f253683-d1a4-48cd-ab57-44cfd7411506 req-9d291b01-a4ee-4c8b-8d83-8a68dab5b91f service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Refreshing network info cache for port dd8cb7f7-fc67-4477-bdd4-27617d60e42e {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1160.145790] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240212, 'name': PowerOffVM_Task, 'duration_secs': 0.180588} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.146085] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1160.146318] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance 'a7f32272-20db-4291-8ef8-bce7f56b6893' progress to 17 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1160.421964] env[62552]: DEBUG nova.network.neutron [req-2f253683-d1a4-48cd-ab57-44cfd7411506 req-9d291b01-a4ee-4c8b-8d83-8a68dab5b91f service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Updated VIF entry in instance network info cache for port dd8cb7f7-fc67-4477-bdd4-27617d60e42e. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1160.422345] env[62552]: DEBUG nova.network.neutron [req-2f253683-d1a4-48cd-ab57-44cfd7411506 req-9d291b01-a4ee-4c8b-8d83-8a68dab5b91f service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Updating instance_info_cache with network_info: [{"id": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "address": "fa:16:3e:75:d0:12", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd8cb7f7-fc", "ovs_interfaceid": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.653025] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1160.653321] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1160.653547] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1160.653803] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1160.654032] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1160.654210] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1160.654499] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1160.654728] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1160.654946] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1160.655199] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1160.655439] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1160.661104] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-806c4799-352c-45d6-9e6e-797186172897 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.678396] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1160.678396] env[62552]: value = "task-1240213" [ 1160.678396] env[62552]: _type = "Task" [ 1160.678396] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.687856] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240213, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.925421] env[62552]: DEBUG oslo_concurrency.lockutils [req-2f253683-d1a4-48cd-ab57-44cfd7411506 req-9d291b01-a4ee-4c8b-8d83-8a68dab5b91f service nova] Releasing lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.190278] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240213, 'name': ReconfigVM_Task, 'duration_secs': 0.153975} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.190717] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance 'a7f32272-20db-4291-8ef8-bce7f56b6893' progress to 33 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1161.239709] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.239872] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquired lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.240039] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Forcefully refreshing network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1161.699127] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1161.699587] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1161.699587] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1161.699683] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1161.699813] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1161.700021] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1161.700272] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1161.700441] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1161.700609] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1161.700779] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1161.700946] env[62552]: DEBUG nova.virt.hardware [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1161.706227] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1161.706525] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8be1cff1-b6c4-4b9f-a377-49ea9827933f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.724801] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1161.724801] env[62552]: value = "task-1240214" [ 1161.724801] env[62552]: _type = "Task" [ 1161.724801] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.732464] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240214, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.237395] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240214, 'name': ReconfigVM_Task, 'duration_secs': 0.162562} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.237727] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1162.238532] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2abbcf-bbb5-4900-8f72-60ae242afb81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.260593] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da/volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.260928] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16e78930-7d5a-47b2-a017-99b6f12e205a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.281428] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1162.281428] env[62552]: value = "task-1240215" [ 1162.281428] env[62552]: _type = "Task" [ 1162.281428] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.289497] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240215, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.448468] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Updating instance_info_cache with network_info: [{"id": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "address": "fa:16:3e:25:df:0e", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae5599ec-5b", "ovs_interfaceid": "ae5599ec-5b26-49c6-86d1-165c7baf1b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.791476] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240215, 'name': ReconfigVM_Task, 'duration_secs': 0.261625} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.791804] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da/volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da.vmdk or device None with type thin {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1162.792028] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance 'a7f32272-20db-4291-8ef8-bce7f56b6893' progress to 50 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1162.950840] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Releasing lock "refresh_cache-888ed8fb-18e6-4f69-a27c-bd24c19e0a12" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.951082] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Updated the network info_cache for instance {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1162.951287] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.951445] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.951611] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.951740] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1162.951911] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.298420] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13712947-5991-4cf7-be7f-a589d9264faa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.318477] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccccc650-4518-470f-8864-fcb8f0a1ef17 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.335170] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance 'a7f32272-20db-4291-8ef8-bce7f56b6893' progress to 67 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1163.455201] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.455417] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.455591] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.455745] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1163.456608] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaad6a34-3b0f-48ce-a737-54829a4d54e1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.464468] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7ab332-4175-4eb6-90b0-ba2f556ab5bd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.478071] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e61e2d2-b9b3-4f8b-915f-6bae33834e09 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.484090] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576a0490-ae27-4289-9eb0-7cff66d98b33 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.513612] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180893MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1163.513752] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.513955] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.521516] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Applying migration context for instance a7f32272-20db-4291-8ef8-bce7f56b6893 as it has an incoming, in-progress migration 02e864b8-cba0-41c6-b216-564df74dfd9e. Migration status is migrating {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1164.522350] env[62552]: INFO nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating resource usage from migration 02e864b8-cba0-41c6-b216-564df74dfd9e [ 1164.540265] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6a5b5f4e-0c84-447d-a3da-8258512abb7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1164.540423] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 888ed8fb-18e6-4f69-a27c-bd24c19e0a12 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1164.540548] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 44b075d0-66ac-4b49-928d-d15db17a282e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1164.540668] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Migration 02e864b8-cba0-41c6-b216-564df74dfd9e is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1164.540805] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance a7f32272-20db-4291-8ef8-bce7f56b6893 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1164.541080] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1164.541252] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1164.604708] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21b232a-3d2d-45af-afeb-30d04aa54c2d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.612238] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6f17c6-b976-4509-b435-6ada37e0d211 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.641210] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c82fc9-2c1b-4724-82c6-ff63687ef9ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.647850] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759f563e-6eb5-445f-b74f-381b09b69b43 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.660163] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1164.975938] env[62552]: DEBUG nova.network.neutron [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Port 1961d456-c180-4993-a8b5-33fab8b7e18c binding to destination host cpu-1 is already ACTIVE {{(pid=62552) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1165.162934] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1165.667373] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1165.667777] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.154s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.997889] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "a7f32272-20db-4291-8ef8-bce7f56b6893-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.998138] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.998305] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.414124] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1167.031571] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.031790] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.032043] env[62552]: DEBUG nova.network.neutron [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1167.731923] env[62552]: DEBUG nova.network.neutron [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance_info_cache with network_info: [{"id": "1961d456-c180-4993-a8b5-33fab8b7e18c", "address": "fa:16:3e:ce:0d:1a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1961d456-c1", "ovs_interfaceid": "1961d456-c180-4993-a8b5-33fab8b7e18c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.234675] env[62552]: DEBUG oslo_concurrency.lockutils [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.743437] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79543b56-1983-4c60-8267-6c9f4089660a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.750636] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de3185f-f040-47b1-9cfe-b62ab442275e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.839171] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c488b715-be9a-4a18-aa3d-3bc0b70753ba {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.859929] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9d65c6-2bae-4717-a725-d4a5ae75cf93 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.866779] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance 'a7f32272-20db-4291-8ef8-bce7f56b6893' progress to 83 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1170.374081] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1170.374081] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1f15684-cf88-43c0-9fa4-1362e17d868c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.381907] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1170.381907] env[62552]: value = "task-1240216" [ 1170.381907] env[62552]: _type = "Task" [ 1170.381907] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.389882] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240216, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.891914] env[62552]: DEBUG oslo_vmware.api [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240216, 'name': PowerOnVM_Task, 'duration_secs': 0.379656} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.892312] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1170.892371] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-082dd5bc-abcc-43a1-a81e-4bae1c55ca4f tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance 'a7f32272-20db-4291-8ef8-bce7f56b6893' progress to 100 {{(pid=62552) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1172.714311] env[62552]: DEBUG oslo_concurrency.lockutils [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "a7f32272-20db-4291-8ef8-bce7f56b6893" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.714980] env[62552]: DEBUG oslo_concurrency.lockutils [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.715220] env[62552]: DEBUG nova.compute.manager [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Going to confirm migration 6 {{(pid=62552) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 1173.278162] env[62552]: DEBUG oslo_concurrency.lockutils [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.278374] env[62552]: DEBUG oslo_concurrency.lockutils [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquired lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.278558] env[62552]: DEBUG nova.network.neutron [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1173.278746] env[62552]: DEBUG nova.objects.instance [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lazy-loading 'info_cache' on Instance uuid a7f32272-20db-4291-8ef8-bce7f56b6893 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1174.475956] env[62552]: DEBUG nova.network.neutron [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance_info_cache with network_info: [{"id": "1961d456-c180-4993-a8b5-33fab8b7e18c", "address": "fa:16:3e:ce:0d:1a", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1961d456-c1", "ovs_interfaceid": "1961d456-c180-4993-a8b5-33fab8b7e18c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.978340] env[62552]: DEBUG oslo_concurrency.lockutils [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Releasing lock "refresh_cache-a7f32272-20db-4291-8ef8-bce7f56b6893" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1174.978650] env[62552]: DEBUG nova.objects.instance [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lazy-loading 'migration_context' on Instance uuid a7f32272-20db-4291-8ef8-bce7f56b6893 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1175.481801] env[62552]: DEBUG nova.objects.base [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1175.482811] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48aa80d9-c01b-4336-a62c-c98f992c2c7f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.502124] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80e8f98f-7ba2-4976-a24d-b6d583a0bb74 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.507326] env[62552]: DEBUG oslo_vmware.api [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1175.507326] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f1889d-ff97-4271-3050-bc3efac8d10c" [ 1175.507326] env[62552]: _type = "Task" [ 1175.507326] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.514770] env[62552]: DEBUG oslo_vmware.api [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f1889d-ff97-4271-3050-bc3efac8d10c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.017991] env[62552]: DEBUG oslo_vmware.api [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52f1889d-ff97-4271-3050-bc3efac8d10c, 'name': SearchDatastore_Task, 'duration_secs': 0.007237} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.019338] env[62552]: DEBUG oslo_concurrency.lockutils [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.019338] env[62552]: DEBUG oslo_concurrency.lockutils [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.587415] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b847afba-8b19-44fa-b711-959853274f3a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.594999] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa1176f-e033-4b83-9355-e76ad01489ec {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.626733] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e00e1fea-024a-446f-8cf4-46102f5305c3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.634255] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c118119-e4a2-4376-88e3-1f3965349653 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.647440] env[62552]: DEBUG nova.compute.provider_tree [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.150461] env[62552]: DEBUG nova.scheduler.client.report [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1178.161341] env[62552]: DEBUG oslo_concurrency.lockutils [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.142s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.712629] env[62552]: INFO nova.scheduler.client.report [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted allocation for migration 02e864b8-cba0-41c6-b216-564df74dfd9e [ 1179.093902] env[62552]: INFO nova.compute.manager [None req-606b1081-a8af-4562-a245-f5d3d01a186e tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Get console output [ 1179.094338] env[62552]: WARNING nova.virt.vmwareapi.driver [None req-606b1081-a8af-4562-a245-f5d3d01a186e tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] The console log is missing. Check your VSPC configuration [ 1179.217941] env[62552]: DEBUG oslo_concurrency.lockutils [None req-edfc0eb5-218a-4b72-a36a-a756f8b63458 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.503s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.252465] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "44b075d0-66ac-4b49-928d-d15db17a282e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.252801] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "44b075d0-66ac-4b49-928d-d15db17a282e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.252945] env[62552]: DEBUG nova.compute.manager [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1196.253861] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712724ae-4605-45f0-ba7a-777a2bcda9f0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.260838] env[62552]: DEBUG nova.compute.manager [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62552) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1196.261910] env[62552]: DEBUG nova.objects.instance [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'flavor' on Instance uuid 44b075d0-66ac-4b49-928d-d15db17a282e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.270025] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1197.270387] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c3325e40-b8b0-4c99-97c1-070d9e87fb9c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.278781] env[62552]: DEBUG oslo_vmware.api [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1197.278781] env[62552]: value = "task-1240217" [ 1197.278781] env[62552]: _type = "Task" [ 1197.278781] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.287623] env[62552]: DEBUG oslo_vmware.api [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240217, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.789681] env[62552]: DEBUG oslo_vmware.api [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240217, 'name': PowerOffVM_Task, 'duration_secs': 0.19414} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.789681] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1197.789921] env[62552]: DEBUG nova.compute.manager [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1197.790578] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364cb602-0859-4232-bb30-44629783b721 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.301692] env[62552]: DEBUG oslo_concurrency.lockutils [None req-3e7d8e38-43cb-42c1-853a-a9e55ab36cab tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "44b075d0-66ac-4b49-928d-d15db17a282e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.049s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.634069] env[62552]: DEBUG nova.objects.instance [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'flavor' on Instance uuid 44b075d0-66ac-4b49-928d-d15db17a282e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.139160] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.139367] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.139524] env[62552]: DEBUG nova.network.neutron [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1199.139704] env[62552]: DEBUG nova.objects.instance [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'info_cache' on Instance uuid 44b075d0-66ac-4b49-928d-d15db17a282e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.643202] env[62552]: DEBUG nova.objects.base [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Object Instance<44b075d0-66ac-4b49-928d-d15db17a282e> lazy-loaded attributes: flavor,info_cache {{(pid=62552) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1200.376167] env[62552]: DEBUG nova.network.neutron [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Updating instance_info_cache with network_info: [{"id": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "address": "fa:16:3e:75:d0:12", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd8cb7f7-fc", "ovs_interfaceid": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.879881] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.886983] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1201.887392] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20990da8-2717-4d92-ae32-d49022fdcd91 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.895012] env[62552]: DEBUG oslo_vmware.api [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1201.895012] env[62552]: value = "task-1240218" [ 1201.895012] env[62552]: _type = "Task" [ 1201.895012] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.902615] env[62552]: DEBUG oslo_vmware.api [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240218, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.405821] env[62552]: DEBUG oslo_vmware.api [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240218, 'name': PowerOnVM_Task, 'duration_secs': 0.383485} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.406180] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1202.406400] env[62552]: DEBUG nova.compute.manager [None req-bae0cfff-83c8-48c2-9694-c9b8c9bb5137 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1202.407280] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab37c010-7c51-4719-95a8-d8b1f1353e1b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.593916] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fdb471-e50b-4f52-9f90-1bbf0812926c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.600669] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5f14e8cd-b3f4-4887-94f7-4e430f6ec98c tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Suspending the VM {{(pid=62552) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1203.600898] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-78579030-5272-40b5-9028-e9976eff9c42 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.607260] env[62552]: DEBUG oslo_vmware.api [None req-5f14e8cd-b3f4-4887-94f7-4e430f6ec98c tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1203.607260] env[62552]: value = "task-1240219" [ 1203.607260] env[62552]: _type = "Task" [ 1203.607260] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.614866] env[62552]: DEBUG oslo_vmware.api [None req-5f14e8cd-b3f4-4887-94f7-4e430f6ec98c tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240219, 'name': SuspendVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.117458] env[62552]: DEBUG oslo_vmware.api [None req-5f14e8cd-b3f4-4887-94f7-4e430f6ec98c tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240219, 'name': SuspendVM_Task} progress is 62%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.618166] env[62552]: DEBUG oslo_vmware.api [None req-5f14e8cd-b3f4-4887-94f7-4e430f6ec98c tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240219, 'name': SuspendVM_Task, 'duration_secs': 0.808879} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.618537] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-5f14e8cd-b3f4-4887-94f7-4e430f6ec98c tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Suspended the VM {{(pid=62552) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1204.618587] env[62552]: DEBUG nova.compute.manager [None req-5f14e8cd-b3f4-4887-94f7-4e430f6ec98c tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1204.619355] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6626dca-e3b8-420f-8986-04f966a9afc3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.968603] env[62552]: INFO nova.compute.manager [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Resuming [ 1205.969371] env[62552]: DEBUG nova.objects.instance [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'flavor' on Instance uuid 44b075d0-66ac-4b49-928d-d15db17a282e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1205.975727] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "a7f32272-20db-4291-8ef8-bce7f56b6893" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.975911] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.976151] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "a7f32272-20db-4291-8ef8-bce7f56b6893-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.976360] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.977065] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.978731] env[62552]: INFO nova.compute.manager [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Terminating instance [ 1206.483211] env[62552]: DEBUG nova.compute.manager [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1206.483456] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1206.483970] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac576517-c7d4-4c5d-b27b-06cb0cfb56df {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.492154] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1206.492154] env[62552]: value = "task-1240220" [ 1206.492154] env[62552]: _type = "Task" [ 1206.492154] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.500597] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240220, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.977778] env[62552]: DEBUG oslo_concurrency.lockutils [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1206.978163] env[62552]: DEBUG oslo_concurrency.lockutils [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquired lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.978163] env[62552]: DEBUG nova.network.neutron [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1207.001685] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240220, 'name': PowerOffVM_Task, 'duration_secs': 0.156976} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.001941] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1207.002119] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Volume detach. Driver type: vmdk {{(pid=62552) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1207.002318] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267535', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'name': 'volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'a7f32272-20db-4291-8ef8-bce7f56b6893', 'attached_at': '2024-10-10T11:41:13.000000', 'detached_at': '', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'serial': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1207.003068] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8bcb20-0925-49f4-8148-8ce304a02631 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.021368] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3972bb8-e565-4987-8898-f823cb7af882 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.028082] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc81966-eac2-4fc9-b5f6-2060ffbd8b52 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.046216] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce935a0-227c-4ac5-8d69-9243c2e84794 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.061963] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] The volume has not been displaced from its original location: [datastore2] volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da/volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da.vmdk. No consolidation needed. {{(pid=62552) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1207.067096] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1207.067361] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87ab742a-45b8-495d-8d87-52210567ea77 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.084056] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1207.084056] env[62552]: value = "task-1240221" [ 1207.084056] env[62552]: _type = "Task" [ 1207.084056] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.091235] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240221, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.595776] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240221, 'name': ReconfigVM_Task, 'duration_secs': 0.154521} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.596077] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62552) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1207.600570] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d3286b5-bdcc-4fe3-bf31-da62dee04b9d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.614542] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1207.614542] env[62552]: value = "task-1240222" [ 1207.614542] env[62552]: _type = "Task" [ 1207.614542] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.624328] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240222, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.687931] env[62552]: DEBUG nova.network.neutron [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Updating instance_info_cache with network_info: [{"id": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "address": "fa:16:3e:75:d0:12", "network": {"id": "4ff4b742-b619-4ba2-85d4-3ab9f7215dda", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1418039526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8495bac65967441996ecca1fd22a4da4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd8cb7f7-fc", "ovs_interfaceid": "dd8cb7f7-fc67-4477-bdd4-27617d60e42e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1208.127011] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240222, 'name': ReconfigVM_Task, 'duration_secs': 0.130484} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.127369] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-267535', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'name': 'volume-b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'a7f32272-20db-4291-8ef8-bce7f56b6893', 'attached_at': '2024-10-10T11:41:13.000000', 'detached_at': '', 'volume_id': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da', 'serial': 'b48a8f1d-dc4a-4b43-ace9-214a64d673da'} {{(pid=62552) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1208.127555] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1208.128307] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753d490b-813c-4f35-b7f2-0dba5dc37497 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.134553] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1208.134769] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1312dd90-197e-4a06-9b4e-118d9f5fd392 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.191030] env[62552]: DEBUG oslo_concurrency.lockutils [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Releasing lock "refresh_cache-44b075d0-66ac-4b49-928d-d15db17a282e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.192037] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f552bf0-08de-495d-a8c8-a78f892bee32 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.196189] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1208.196398] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1208.196582] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleting the datastore file [datastore2] a7f32272-20db-4291-8ef8-bce7f56b6893 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1208.197219] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fc7b8de-7975-4702-96b8-50f39fe78e1a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.200479] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Resuming the VM {{(pid=62552) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1208.200695] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d8e7e67-157b-4638-b35b-950fd8a9e46f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.202898] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1208.202898] env[62552]: value = "task-1240224" [ 1208.202898] env[62552]: _type = "Task" [ 1208.202898] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.206749] env[62552]: DEBUG oslo_vmware.api [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1208.206749] env[62552]: value = "task-1240225" [ 1208.206749] env[62552]: _type = "Task" [ 1208.206749] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.212697] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240224, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.216928] env[62552]: DEBUG oslo_vmware.api [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240225, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.716596] env[62552]: DEBUG oslo_vmware.api [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240225, 'name': PowerOnVM_Task, 'duration_secs': 0.490142} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.719759] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Resumed the VM {{(pid=62552) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1208.719996] env[62552]: DEBUG nova.compute.manager [None req-708ed527-81fc-4e01-b6d1-49cc1027874d tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1208.720311] env[62552]: DEBUG oslo_vmware.api [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240224, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.08179} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.721046] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0ee221-98d9-48bd-9f38-b711d277f026 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.723625] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1208.723850] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1208.724065] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1208.724208] env[62552]: INFO nova.compute.manager [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Took 2.24 seconds to destroy the instance on the hypervisor. [ 1208.724455] env[62552]: DEBUG oslo.service.loopingcall [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1208.724693] env[62552]: DEBUG nova.compute.manager [-] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1208.724794] env[62552]: DEBUG nova.network.neutron [-] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1209.194981] env[62552]: DEBUG nova.compute.manager [req-fa966ad3-271a-40f2-b1fe-37819e2d2bfc req-ec5f7c69-cf9a-4b1e-85f7-75f7c503e487 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Received event network-vif-deleted-1961d456-c180-4993-a8b5-33fab8b7e18c {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1209.195297] env[62552]: INFO nova.compute.manager [req-fa966ad3-271a-40f2-b1fe-37819e2d2bfc req-ec5f7c69-cf9a-4b1e-85f7-75f7c503e487 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Neutron deleted interface 1961d456-c180-4993-a8b5-33fab8b7e18c; detaching it from the instance and deleting it from the info cache [ 1209.195297] env[62552]: DEBUG nova.network.neutron [req-fa966ad3-271a-40f2-b1fe-37819e2d2bfc req-ec5f7c69-cf9a-4b1e-85f7-75f7c503e487 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.673750] env[62552]: DEBUG nova.network.neutron [-] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.697926] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f2c5e66-b003-4408-b7a2-f799f2360ab8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.707406] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f1d548-e7fc-4af9-9be4-b27f7a8cfef8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.717742] env[62552]: DEBUG oslo_concurrency.lockutils [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "44b075d0-66ac-4b49-928d-d15db17a282e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.718159] env[62552]: DEBUG oslo_concurrency.lockutils [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "44b075d0-66ac-4b49-928d-d15db17a282e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.718416] env[62552]: DEBUG oslo_concurrency.lockutils [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "44b075d0-66ac-4b49-928d-d15db17a282e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.718644] env[62552]: DEBUG oslo_concurrency.lockutils [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "44b075d0-66ac-4b49-928d-d15db17a282e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.718827] env[62552]: DEBUG oslo_concurrency.lockutils [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "44b075d0-66ac-4b49-928d-d15db17a282e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.721109] env[62552]: INFO nova.compute.manager [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Terminating instance [ 1209.734589] env[62552]: DEBUG nova.compute.manager [req-fa966ad3-271a-40f2-b1fe-37819e2d2bfc req-ec5f7c69-cf9a-4b1e-85f7-75f7c503e487 service nova] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Detach interface failed, port_id=1961d456-c180-4993-a8b5-33fab8b7e18c, reason: Instance a7f32272-20db-4291-8ef8-bce7f56b6893 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1210.176907] env[62552]: INFO nova.compute.manager [-] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Took 1.45 seconds to deallocate network for instance. [ 1210.225205] env[62552]: DEBUG nova.compute.manager [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1210.225510] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1210.226406] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9bb54f-629a-49dc-a23a-120fa230753f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.234451] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1210.234709] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-81acf536-3904-49ca-887f-cee5e7e1691e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.241698] env[62552]: DEBUG oslo_vmware.api [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1210.241698] env[62552]: value = "task-1240226" [ 1210.241698] env[62552]: _type = "Task" [ 1210.241698] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.249581] env[62552]: DEBUG oslo_vmware.api [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240226, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.721583] env[62552]: INFO nova.compute.manager [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Took 0.54 seconds to detach 1 volumes for instance. [ 1210.723826] env[62552]: DEBUG nova.compute.manager [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: a7f32272-20db-4291-8ef8-bce7f56b6893] Deleting volume: b48a8f1d-dc4a-4b43-ace9-214a64d673da {{(pid=62552) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 1210.753254] env[62552]: DEBUG oslo_vmware.api [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240226, 'name': PowerOffVM_Task, 'duration_secs': 0.183364} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.753517] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1210.753690] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1210.753926] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bbf96881-231b-4e58-a259-994159e75b6b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.821549] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1210.821732] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1210.821919] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleting the datastore file [datastore2] 44b075d0-66ac-4b49-928d-d15db17a282e {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1210.822210] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8cf7282-e0bd-49cd-b79e-7ab704e39c7b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.828433] env[62552]: DEBUG oslo_vmware.api [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for the task: (returnval){ [ 1210.828433] env[62552]: value = "task-1240229" [ 1210.828433] env[62552]: _type = "Task" [ 1210.828433] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.835535] env[62552]: DEBUG oslo_vmware.api [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240229, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.260539] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.260890] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.261023] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.285270] env[62552]: INFO nova.scheduler.client.report [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted allocations for instance a7f32272-20db-4291-8ef8-bce7f56b6893 [ 1211.337986] env[62552]: DEBUG oslo_vmware.api [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Task: {'id': task-1240229, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151867} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.340529] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1211.340529] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1211.340529] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1211.340529] env[62552]: INFO nova.compute.manager [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1211.340529] env[62552]: DEBUG oslo.service.loopingcall [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1211.340529] env[62552]: DEBUG nova.compute.manager [-] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1211.340529] env[62552]: DEBUG nova.network.neutron [-] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1211.777376] env[62552]: DEBUG nova.compute.manager [req-c88f0f53-331f-4647-9a1c-ba863b37c1a8 req-b78f0f76-3209-4c4f-89eb-7d8fc50761d3 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Received event network-vif-deleted-dd8cb7f7-fc67-4477-bdd4-27617d60e42e {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1211.777598] env[62552]: INFO nova.compute.manager [req-c88f0f53-331f-4647-9a1c-ba863b37c1a8 req-b78f0f76-3209-4c4f-89eb-7d8fc50761d3 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Neutron deleted interface dd8cb7f7-fc67-4477-bdd4-27617d60e42e; detaching it from the instance and deleting it from the info cache [ 1211.777749] env[62552]: DEBUG nova.network.neutron [req-c88f0f53-331f-4647-9a1c-ba863b37c1a8 req-b78f0f76-3209-4c4f-89eb-7d8fc50761d3 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.793461] env[62552]: DEBUG oslo_concurrency.lockutils [None req-bdd1a21f-32ce-4c9e-91e2-5a8aeda12794 tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "a7f32272-20db-4291-8ef8-bce7f56b6893" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.817s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.252486] env[62552]: DEBUG nova.network.neutron [-] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.280517] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5590f70a-1808-49b2-9af2-72cf9d6dfcf0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.291268] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f647bbf4-e24e-454e-9e7c-99098d225e04 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.315871] env[62552]: DEBUG nova.compute.manager [req-c88f0f53-331f-4647-9a1c-ba863b37c1a8 req-b78f0f76-3209-4c4f-89eb-7d8fc50761d3 service nova] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Detach interface failed, port_id=dd8cb7f7-fc67-4477-bdd4-27617d60e42e, reason: Instance 44b075d0-66ac-4b49-928d-d15db17a282e could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1212.755828] env[62552]: INFO nova.compute.manager [-] [instance: 44b075d0-66ac-4b49-928d-d15db17a282e] Took 1.42 seconds to deallocate network for instance. [ 1212.763882] env[62552]: DEBUG oslo_concurrency.lockutils [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.764125] env[62552]: DEBUG oslo_concurrency.lockutils [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.764322] env[62552]: DEBUG oslo_concurrency.lockutils [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.764508] env[62552]: DEBUG oslo_concurrency.lockutils [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.764728] env[62552]: DEBUG oslo_concurrency.lockutils [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.766359] env[62552]: INFO nova.compute.manager [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Terminating instance [ 1213.262512] env[62552]: DEBUG oslo_concurrency.lockutils [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.262866] env[62552]: DEBUG oslo_concurrency.lockutils [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.263176] env[62552]: DEBUG nova.objects.instance [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lazy-loading 'resources' on Instance uuid 44b075d0-66ac-4b49-928d-d15db17a282e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.272468] env[62552]: DEBUG nova.compute.manager [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1213.272706] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1213.273579] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9f86d9-b40a-49e2-8f7e-eee493bff9e6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.281585] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1213.281865] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6055b20c-a3b7-4a11-9b6b-00fdfa907d92 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.289494] env[62552]: DEBUG oslo_vmware.api [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1213.289494] env[62552]: value = "task-1240230" [ 1213.289494] env[62552]: _type = "Task" [ 1213.289494] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.296891] env[62552]: DEBUG oslo_vmware.api [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240230, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.801929] env[62552]: DEBUG oslo_vmware.api [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240230, 'name': PowerOffVM_Task, 'duration_secs': 0.21229} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.802226] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1213.802406] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1213.802656] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae32b3b1-9815-4fc2-80e8-d329564888c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.814426] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2e4b47-cfd2-4949-a420-06962abe508a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.821112] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6faec0d4-6e97-4002-b478-1f86ac08d113 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.851810] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ba405b-05ad-4eed-8c74-2c1f090ecda4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.859406] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138e01bb-30b6-4c98-802b-e5d8e9d61cdb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.872468] env[62552]: DEBUG nova.compute.provider_tree [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1213.874582] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1213.874771] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1213.874957] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleting the datastore file [datastore2] 888ed8fb-18e6-4f69-a27c-bd24c19e0a12 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1213.875405] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-923dd3a1-d48f-46df-baf0-448f408183fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.881105] env[62552]: DEBUG oslo_vmware.api [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1213.881105] env[62552]: value = "task-1240232" [ 1213.881105] env[62552]: _type = "Task" [ 1213.881105] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.889639] env[62552]: DEBUG oslo_vmware.api [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240232, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.376509] env[62552]: DEBUG nova.scheduler.client.report [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1214.390831] env[62552]: DEBUG oslo_vmware.api [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240232, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129782} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.391542] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1214.391736] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1214.391944] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1214.392135] env[62552]: INFO nova.compute.manager [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1214.392380] env[62552]: DEBUG oslo.service.loopingcall [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1214.392596] env[62552]: DEBUG nova.compute.manager [-] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1214.392697] env[62552]: DEBUG nova.network.neutron [-] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1214.650418] env[62552]: DEBUG nova.compute.manager [req-775a25c1-a984-4626-9820-0d7e7d8ebec9 req-f2d09daa-6f96-4137-83fc-a40d8b952965 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Received event network-vif-deleted-ae5599ec-5b26-49c6-86d1-165c7baf1b2f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1214.650622] env[62552]: INFO nova.compute.manager [req-775a25c1-a984-4626-9820-0d7e7d8ebec9 req-f2d09daa-6f96-4137-83fc-a40d8b952965 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Neutron deleted interface ae5599ec-5b26-49c6-86d1-165c7baf1b2f; detaching it from the instance and deleting it from the info cache [ 1214.650810] env[62552]: DEBUG nova.network.neutron [req-775a25c1-a984-4626-9820-0d7e7d8ebec9 req-f2d09daa-6f96-4137-83fc-a40d8b952965 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.881511] env[62552]: DEBUG oslo_concurrency.lockutils [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.902257] env[62552]: INFO nova.scheduler.client.report [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Deleted allocations for instance 44b075d0-66ac-4b49-928d-d15db17a282e [ 1215.131616] env[62552]: DEBUG nova.network.neutron [-] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1215.153453] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-004a46a5-3023-4a59-891e-feaecca1da69 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.165326] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6867d81-c280-4fa1-856c-0b0193c576af {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.187714] env[62552]: DEBUG nova.compute.manager [req-775a25c1-a984-4626-9820-0d7e7d8ebec9 req-f2d09daa-6f96-4137-83fc-a40d8b952965 service nova] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Detach interface failed, port_id=ae5599ec-5b26-49c6-86d1-165c7baf1b2f, reason: Instance 888ed8fb-18e6-4f69-a27c-bd24c19e0a12 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1215.410700] env[62552]: DEBUG oslo_concurrency.lockutils [None req-02d2e238-dd09-404a-8757-7d95108a1700 tempest-ServerActionsTestJSON-209267675 tempest-ServerActionsTestJSON-209267675-project-member] Lock "44b075d0-66ac-4b49-928d-d15db17a282e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.692s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.634602] env[62552]: INFO nova.compute.manager [-] [instance: 888ed8fb-18e6-4f69-a27c-bd24c19e0a12] Took 1.24 seconds to deallocate network for instance. [ 1216.141728] env[62552]: DEBUG oslo_concurrency.lockutils [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.142062] env[62552]: DEBUG oslo_concurrency.lockutils [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.142398] env[62552]: DEBUG nova.objects.instance [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lazy-loading 'resources' on Instance uuid 888ed8fb-18e6-4f69-a27c-bd24c19e0a12 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.687074] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806fb619-30a9-46db-ae53-03f17b733705 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.695034] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa201b9d-c2fe-4223-b1d0-79030d4aa3db {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.723799] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce54c0e3-0436-4801-bc97-ee0fb209ae35 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.730698] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8704042a-9e44-4bfb-8979-21a183c9215c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.746803] env[62552]: DEBUG nova.compute.provider_tree [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.250784] env[62552]: DEBUG nova.scheduler.client.report [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1217.755842] env[62552]: DEBUG oslo_concurrency.lockutils [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.614s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.779834] env[62552]: INFO nova.scheduler.client.report [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted allocations for instance 888ed8fb-18e6-4f69-a27c-bd24c19e0a12 [ 1218.288636] env[62552]: DEBUG oslo_concurrency.lockutils [None req-172da861-4396-45fb-8998-49b14d8d02cc tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "888ed8fb-18e6-4f69-a27c-bd24c19e0a12" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.524s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.697995] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.698270] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.693578] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.697300] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.697573] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.697898] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1220.697898] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Rebuilding the list of instances to heal {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1220.752275] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquiring lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.752575] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1221.232419] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1221.232599] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquired lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.232787] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Forcefully refreshing network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1221.232953] env[62552]: DEBUG nova.objects.instance [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lazy-loading 'info_cache' on Instance uuid 6a5b5f4e-0c84-447d-a3da-8258512abb7c {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1221.254442] env[62552]: DEBUG nova.compute.manager [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1221.776132] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1221.776426] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1221.777836] env[62552]: INFO nova.compute.claims [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1222.590153] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.590445] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.590656] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.590845] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.591031] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.593226] env[62552]: INFO nova.compute.manager [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Terminating instance [ 1222.820443] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393655e0-9242-4b72-9951-7648e8f363e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.830044] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db291d2-cbdb-42dc-a3ba-364fffd7eb55 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.861774] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4c3a76-6663-42c1-93ec-dd4f07763976 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.869478] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f8d42e-5e99-439a-bd7a-6ee257d9d6d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.882320] env[62552]: DEBUG nova.compute.provider_tree [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1222.956539] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updating instance_info_cache with network_info: [{"id": "5e69a23a-a63b-4d24-9729-859b537f945b", "address": "fa:16:3e:0b:c1:98", "network": {"id": "f915d0b6-087b-4e19-9c2b-9340eb497579", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-489250545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a0bc011d6794602b2bbe1fc01e4c8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e69a23a-a6", "ovs_interfaceid": "5e69a23a-a63b-4d24-9729-859b537f945b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1223.096879] env[62552]: DEBUG nova.compute.manager [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1223.097127] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1223.098015] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff89fcb-c4d9-41a8-9715-48361c2955e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.107061] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1223.107288] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10dfd148-46f8-4acd-9f1d-bafcde1687c1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.113372] env[62552]: DEBUG oslo_vmware.api [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1223.113372] env[62552]: value = "task-1240234" [ 1223.113372] env[62552]: _type = "Task" [ 1223.113372] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.121453] env[62552]: DEBUG oslo_vmware.api [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.385231] env[62552]: DEBUG nova.scheduler.client.report [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1223.459173] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Releasing lock "refresh_cache-6a5b5f4e-0c84-447d-a3da-8258512abb7c" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1223.459393] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updated the network info_cache for instance {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1223.459598] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1223.459771] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1223.459903] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1223.460067] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1223.623273] env[62552]: DEBUG oslo_vmware.api [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240234, 'name': PowerOffVM_Task, 'duration_secs': 0.184112} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.623506] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1223.623680] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1223.623929] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8fce3c4-3e78-4168-8774-e6ef979d5973 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.683827] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1223.684072] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1223.684263] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleting the datastore file [datastore2] 6a5b5f4e-0c84-447d-a3da-8258512abb7c {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1223.684531] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f474ae2-3703-4901-8c98-fd4d86d4ab09 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.690031] env[62552]: DEBUG oslo_vmware.api [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for the task: (returnval){ [ 1223.690031] env[62552]: value = "task-1240236" [ 1223.690031] env[62552]: _type = "Task" [ 1223.690031] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.697236] env[62552]: DEBUG oslo_vmware.api [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240236, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.891069] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.114s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.891595] env[62552]: DEBUG nova.compute.manager [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1223.962582] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.962921] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.963201] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.963433] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1223.965007] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ed185c-9aaa-445e-9560-bc7c324fcd9b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.973866] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74226529-acd9-4492-85b3-98a75b86e5a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.987404] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df93ada-0934-49e7-9d64-ae4acaf39c08 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.993551] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7248fec7-aa83-49d1-b292-ac10aa6911a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.022251] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180773MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1224.022390] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.022593] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.201523] env[62552]: DEBUG oslo_vmware.api [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Task: {'id': task-1240236, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129371} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.201956] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1224.202278] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1224.202572] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1224.202867] env[62552]: INFO nova.compute.manager [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1224.203280] env[62552]: DEBUG oslo.service.loopingcall [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1224.203573] env[62552]: DEBUG nova.compute.manager [-] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1224.203725] env[62552]: DEBUG nova.network.neutron [-] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1224.396866] env[62552]: DEBUG nova.compute.utils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1224.398634] env[62552]: DEBUG nova.compute.manager [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1224.398915] env[62552]: DEBUG nova.network.neutron [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1224.485697] env[62552]: DEBUG nova.compute.manager [req-ef75deea-bd5d-41b2-bcf8-0c3168776a86 req-8e0bbabb-0723-4eaa-adee-172ed2288704 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Received event network-vif-deleted-5e69a23a-a63b-4d24-9729-859b537f945b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1224.485808] env[62552]: INFO nova.compute.manager [req-ef75deea-bd5d-41b2-bcf8-0c3168776a86 req-8e0bbabb-0723-4eaa-adee-172ed2288704 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Neutron deleted interface 5e69a23a-a63b-4d24-9729-859b537f945b; detaching it from the instance and deleting it from the info cache [ 1224.485988] env[62552]: DEBUG nova.network.neutron [req-ef75deea-bd5d-41b2-bcf8-0c3168776a86 req-8e0bbabb-0723-4eaa-adee-172ed2288704 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.491092] env[62552]: DEBUG nova.policy [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ada89037f20c4bb29ed7616d8a23e1a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0547139d1307408ebd9defb3cdc44121', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1224.731913] env[62552]: DEBUG nova.network.neutron [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Successfully created port: 37fd08dc-ff04-4070-858d-c78004db354d {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1224.902854] env[62552]: DEBUG nova.compute.manager [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1224.963579] env[62552]: DEBUG nova.network.neutron [-] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.987942] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-095c6865-3081-4730-aba6-a55ef2daf79e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.998012] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6293da5b-b1ad-4ddf-915f-10787837af35 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.019759] env[62552]: DEBUG nova.compute.manager [req-ef75deea-bd5d-41b2-bcf8-0c3168776a86 req-8e0bbabb-0723-4eaa-adee-172ed2288704 service nova] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Detach interface failed, port_id=5e69a23a-a63b-4d24-9729-859b537f945b, reason: Instance 6a5b5f4e-0c84-447d-a3da-8258512abb7c could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1225.045050] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 6a5b5f4e-0c84-447d-a3da-8258512abb7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1225.045210] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance d5d0b737-974a-4b5d-9a8e-01d101f5d476 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1225.045388] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1225.045530] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1225.078876] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bda73b3-225e-434e-9668-6b1254c1bd2c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.086225] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8012d1-36c3-47d6-be69-ef6d9ff4724a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.116246] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8451b21-77e6-42fa-9ae8-1d54409cf777 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.122798] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0da631b-2975-41a0-8659-52ef49f1fac9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.135389] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1225.466934] env[62552]: INFO nova.compute.manager [-] [instance: 6a5b5f4e-0c84-447d-a3da-8258512abb7c] Took 1.26 seconds to deallocate network for instance. [ 1225.638373] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1225.911848] env[62552]: DEBUG nova.compute.manager [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1225.936315] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1225.936556] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1225.936715] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1225.936901] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1225.937073] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1225.937222] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1225.937431] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1225.937593] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1225.937761] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1225.937927] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1225.938113] env[62552]: DEBUG nova.virt.hardware [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1225.938959] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe401e94-105a-40d4-98de-466bd23ebed3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.946845] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c705a908-539a-4a3e-a331-ecd523c01fbb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.972979] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.104614] env[62552]: DEBUG nova.compute.manager [req-e8788052-097f-47a3-89ee-1a463e992afd req-71771e81-ea37-4ce8-8023-310a991b02aa service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Received event network-vif-plugged-37fd08dc-ff04-4070-858d-c78004db354d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1226.104932] env[62552]: DEBUG oslo_concurrency.lockutils [req-e8788052-097f-47a3-89ee-1a463e992afd req-71771e81-ea37-4ce8-8023-310a991b02aa service nova] Acquiring lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.105281] env[62552]: DEBUG oslo_concurrency.lockutils [req-e8788052-097f-47a3-89ee-1a463e992afd req-71771e81-ea37-4ce8-8023-310a991b02aa service nova] Lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.105483] env[62552]: DEBUG oslo_concurrency.lockutils [req-e8788052-097f-47a3-89ee-1a463e992afd req-71771e81-ea37-4ce8-8023-310a991b02aa service nova] Lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.105662] env[62552]: DEBUG nova.compute.manager [req-e8788052-097f-47a3-89ee-1a463e992afd req-71771e81-ea37-4ce8-8023-310a991b02aa service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] No waiting events found dispatching network-vif-plugged-37fd08dc-ff04-4070-858d-c78004db354d {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1226.105862] env[62552]: WARNING nova.compute.manager [req-e8788052-097f-47a3-89ee-1a463e992afd req-71771e81-ea37-4ce8-8023-310a991b02aa service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Received unexpected event network-vif-plugged-37fd08dc-ff04-4070-858d-c78004db354d for instance with vm_state building and task_state spawning. [ 1226.143225] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1226.143413] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.121s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.143668] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.171s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.143890] env[62552]: DEBUG nova.objects.instance [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lazy-loading 'resources' on Instance uuid 6a5b5f4e-0c84-447d-a3da-8258512abb7c {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1226.192104] env[62552]: DEBUG nova.network.neutron [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Successfully updated port: 37fd08dc-ff04-4070-858d-c78004db354d {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1226.681849] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e63a72-0dc1-44ab-bf5f-6191dfb40eb0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.689429] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c7c049-6d1d-4896-a7c1-b2539bb23108 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.694474] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquiring lock "refresh_cache-d5d0b737-974a-4b5d-9a8e-01d101f5d476" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1226.694745] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquired lock "refresh_cache-d5d0b737-974a-4b5d-9a8e-01d101f5d476" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.694818] env[62552]: DEBUG nova.network.neutron [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1226.722412] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d352a7-95b5-4c77-b304-2a702d9893a0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.730099] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a2f8ea-8952-49ef-8384-993e697c9140 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.743575] env[62552]: DEBUG nova.compute.provider_tree [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1227.226039] env[62552]: DEBUG nova.network.neutron [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1227.246288] env[62552]: DEBUG nova.scheduler.client.report [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1227.345030] env[62552]: DEBUG nova.network.neutron [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Updating instance_info_cache with network_info: [{"id": "37fd08dc-ff04-4070-858d-c78004db354d", "address": "fa:16:3e:dd:12:44", "network": {"id": "24d59c8e-854d-4206-a140-3d45f3e3f595", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-189403678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0547139d1307408ebd9defb3cdc44121", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fd08dc-ff", "ovs_interfaceid": "37fd08dc-ff04-4070-858d-c78004db354d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1227.751277] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.773057] env[62552]: INFO nova.scheduler.client.report [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Deleted allocations for instance 6a5b5f4e-0c84-447d-a3da-8258512abb7c [ 1227.847445] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Releasing lock "refresh_cache-d5d0b737-974a-4b5d-9a8e-01d101f5d476" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.847880] env[62552]: DEBUG nova.compute.manager [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Instance network_info: |[{"id": "37fd08dc-ff04-4070-858d-c78004db354d", "address": "fa:16:3e:dd:12:44", "network": {"id": "24d59c8e-854d-4206-a140-3d45f3e3f595", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-189403678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0547139d1307408ebd9defb3cdc44121", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fd08dc-ff", "ovs_interfaceid": "37fd08dc-ff04-4070-858d-c78004db354d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1227.848341] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:12:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73eeba7c-29e1-4fdf-82b3-d62e63e86051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37fd08dc-ff04-4070-858d-c78004db354d', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1227.855725] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Creating folder: Project (0547139d1307408ebd9defb3cdc44121). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1227.856227] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a98619a-d199-461c-b528-b8fe754497d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.867542] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Created folder: Project (0547139d1307408ebd9defb3cdc44121) in parent group-v267339. [ 1227.867735] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Creating folder: Instances. Parent ref: group-v267538. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1227.867961] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c702056-ead1-4563-bf15-da04a51f941e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.875933] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Created folder: Instances in parent group-v267538. [ 1227.876178] env[62552]: DEBUG oslo.service.loopingcall [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1227.876361] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1227.876551] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-442ff5e1-a2f7-4762-910d-93118a49f65c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.894847] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1227.894847] env[62552]: value = "task-1240239" [ 1227.894847] env[62552]: _type = "Task" [ 1227.894847] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.902092] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240239, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.132519] env[62552]: DEBUG nova.compute.manager [req-45b9b1d3-aecd-434d-8af4-7dfb638522cb req-b96ae4ae-86e2-48ed-9cf8-0cdc76099a69 service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Received event network-changed-37fd08dc-ff04-4070-858d-c78004db354d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1228.132764] env[62552]: DEBUG nova.compute.manager [req-45b9b1d3-aecd-434d-8af4-7dfb638522cb req-b96ae4ae-86e2-48ed-9cf8-0cdc76099a69 service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Refreshing instance network info cache due to event network-changed-37fd08dc-ff04-4070-858d-c78004db354d. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1228.132991] env[62552]: DEBUG oslo_concurrency.lockutils [req-45b9b1d3-aecd-434d-8af4-7dfb638522cb req-b96ae4ae-86e2-48ed-9cf8-0cdc76099a69 service nova] Acquiring lock "refresh_cache-d5d0b737-974a-4b5d-9a8e-01d101f5d476" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1228.133155] env[62552]: DEBUG oslo_concurrency.lockutils [req-45b9b1d3-aecd-434d-8af4-7dfb638522cb req-b96ae4ae-86e2-48ed-9cf8-0cdc76099a69 service nova] Acquired lock "refresh_cache-d5d0b737-974a-4b5d-9a8e-01d101f5d476" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.133321] env[62552]: DEBUG nova.network.neutron [req-45b9b1d3-aecd-434d-8af4-7dfb638522cb req-b96ae4ae-86e2-48ed-9cf8-0cdc76099a69 service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Refreshing network info cache for port 37fd08dc-ff04-4070-858d-c78004db354d {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1228.281070] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8053748e-a013-432f-b487-57421bc7530d tempest-ServerActionsTestOtherA-1297073477 tempest-ServerActionsTestOtherA-1297073477-project-member] Lock "6a5b5f4e-0c84-447d-a3da-8258512abb7c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.690s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1228.383290] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1228.404776] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240239, 'name': CreateVM_Task, 'duration_secs': 0.276196} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.404939] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1228.405600] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1228.405802] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.406154] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1228.406394] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46682278-46a8-4f58-96df-372074363c82 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.410661] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1228.410661] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52af305c-72ea-3d8c-ddaf-2b54ff54c044" [ 1228.410661] env[62552]: _type = "Task" [ 1228.410661] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.418219] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52af305c-72ea-3d8c-ddaf-2b54ff54c044, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.840546] env[62552]: DEBUG nova.network.neutron [req-45b9b1d3-aecd-434d-8af4-7dfb638522cb req-b96ae4ae-86e2-48ed-9cf8-0cdc76099a69 service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Updated VIF entry in instance network info cache for port 37fd08dc-ff04-4070-858d-c78004db354d. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1228.840989] env[62552]: DEBUG nova.network.neutron [req-45b9b1d3-aecd-434d-8af4-7dfb638522cb req-b96ae4ae-86e2-48ed-9cf8-0cdc76099a69 service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Updating instance_info_cache with network_info: [{"id": "37fd08dc-ff04-4070-858d-c78004db354d", "address": "fa:16:3e:dd:12:44", "network": {"id": "24d59c8e-854d-4206-a140-3d45f3e3f595", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-189403678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0547139d1307408ebd9defb3cdc44121", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fd08dc-ff", "ovs_interfaceid": "37fd08dc-ff04-4070-858d-c78004db354d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1228.887163] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1228.921325] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52af305c-72ea-3d8c-ddaf-2b54ff54c044, 'name': SearchDatastore_Task, 'duration_secs': 0.008581} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.921708] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1228.921890] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1228.922114] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1228.922271] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.922458] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1228.922774] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-884bcf19-a817-42eb-9639-9f30e9ddb0ea {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.930692] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1228.930881] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1228.931597] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-feff962d-ae1d-4408-9574-32d570456ed9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.936440] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1228.936440] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528f226a-1d4e-5fe8-8671-652c077043a2" [ 1228.936440] env[62552]: _type = "Task" [ 1228.936440] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.943942] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528f226a-1d4e-5fe8-8671-652c077043a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.343663] env[62552]: DEBUG oslo_concurrency.lockutils [req-45b9b1d3-aecd-434d-8af4-7dfb638522cb req-b96ae4ae-86e2-48ed-9cf8-0cdc76099a69 service nova] Releasing lock "refresh_cache-d5d0b737-974a-4b5d-9a8e-01d101f5d476" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1229.447225] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]528f226a-1d4e-5fe8-8671-652c077043a2, 'name': SearchDatastore_Task, 'duration_secs': 0.007773} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.447968] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8407105b-0f63-4531-bff1-fae2f8ee6fe2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.453235] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1229.453235] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b8170d-71ca-8440-8ff1-2892d1e279d4" [ 1229.453235] env[62552]: _type = "Task" [ 1229.453235] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.460289] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b8170d-71ca-8440-8ff1-2892d1e279d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.964293] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52b8170d-71ca-8440-8ff1-2892d1e279d4, 'name': SearchDatastore_Task, 'duration_secs': 0.00886} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.964293] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1229.964293] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] d5d0b737-974a-4b5d-9a8e-01d101f5d476/d5d0b737-974a-4b5d-9a8e-01d101f5d476.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1229.964293] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7400af4-b300-487e-a321-f544a7123351 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.971942] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1229.971942] env[62552]: value = "task-1240240" [ 1229.971942] env[62552]: _type = "Task" [ 1229.971942] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.978813] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240240, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.480752] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240240, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450329} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.481123] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] d5d0b737-974a-4b5d-9a8e-01d101f5d476/d5d0b737-974a-4b5d-9a8e-01d101f5d476.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1230.481363] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1230.481615] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21c33110-9d2a-4598-b147-fde77d86853c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.487585] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1230.487585] env[62552]: value = "task-1240241" [ 1230.487585] env[62552]: _type = "Task" [ 1230.487585] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.498821] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.998315] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064333} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.998405] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1230.999191] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdc93948-2fab-49cc-8ae1-9b4df807a185 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.020926] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] d5d0b737-974a-4b5d-9a8e-01d101f5d476/d5d0b737-974a-4b5d-9a8e-01d101f5d476.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1231.021232] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1dae527e-429b-49f1-aa70-b75bd1b4cfde {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.041532] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1231.041532] env[62552]: value = "task-1240242" [ 1231.041532] env[62552]: _type = "Task" [ 1231.041532] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.049245] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240242, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.551939] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240242, 'name': ReconfigVM_Task, 'duration_secs': 0.352414} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.552271] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Reconfigured VM instance instance-0000006e to attach disk [datastore2] d5d0b737-974a-4b5d-9a8e-01d101f5d476/d5d0b737-974a-4b5d-9a8e-01d101f5d476.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1231.552847] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f7e9053-64de-4d62-a2a4-7f9d9ade4072 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.559805] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1231.559805] env[62552]: value = "task-1240243" [ 1231.559805] env[62552]: _type = "Task" [ 1231.559805] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.568817] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240243, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.069267] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240243, 'name': Rename_Task, 'duration_secs': 0.141668} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.069540] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1232.069786] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b519e8a7-b2a2-46ce-945d-66763f764b27 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.075718] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1232.075718] env[62552]: value = "task-1240244" [ 1232.075718] env[62552]: _type = "Task" [ 1232.075718] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.083794] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240244, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.592256] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240244, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.087991] env[62552]: DEBUG oslo_vmware.api [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240244, 'name': PowerOnVM_Task, 'duration_secs': 0.543112} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.087991] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1233.087991] env[62552]: INFO nova.compute.manager [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Took 7.18 seconds to spawn the instance on the hypervisor. [ 1233.087991] env[62552]: DEBUG nova.compute.manager [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1233.088706] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b6a4f2-d98e-4279-b9b5-6878fc104634 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.606244] env[62552]: INFO nova.compute.manager [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Took 11.85 seconds to build instance. [ 1233.840733] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "604e8472-7af4-49a1-9a0c-459482db9d6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.840951] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "604e8472-7af4-49a1-9a0c-459482db9d6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.108245] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e17f5a3b-da9e-4571-aaa7-572241f29a9e tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.355s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.343259] env[62552]: DEBUG nova.compute.manager [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1234.869131] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.869397] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.870874] env[62552]: INFO nova.compute.claims [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1234.895789] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquiring lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.895897] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.896115] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquiring lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.896315] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.896491] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.898795] env[62552]: INFO nova.compute.manager [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Terminating instance [ 1235.408588] env[62552]: DEBUG nova.compute.manager [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1235.408588] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1235.409331] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92e69f9-5d67-4b2f-b364-44c342821fb0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.417707] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1235.417939] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ab8d9be-1ac7-4c4e-9cfa-a4e7c1bbfad0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.424337] env[62552]: DEBUG oslo_vmware.api [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1235.424337] env[62552]: value = "task-1240245" [ 1235.424337] env[62552]: _type = "Task" [ 1235.424337] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.431556] env[62552]: DEBUG oslo_vmware.api [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.916767] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10cef529-ed3b-423c-9ee9-4ba68a237e8b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.924492] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25042178-ca02-4af3-9f8a-c81c535a66a7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.934964] env[62552]: DEBUG oslo_vmware.api [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240245, 'name': PowerOffVM_Task, 'duration_secs': 0.177616} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.959260] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1235.959449] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1235.959841] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d217f2e7-fd99-46ba-aa94-f7a3caa4adf6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.961708] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4170428-93c4-4efa-bf83-de6f1ec5b7f5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.968608] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a68c7d-862b-4f78-87a7-2606ad1f7526 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.981895] env[62552]: DEBUG nova.compute.provider_tree [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1236.024117] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1236.024375] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1236.024563] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Deleting the datastore file [datastore2] d5d0b737-974a-4b5d-9a8e-01d101f5d476 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1236.024861] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93b93dd0-5731-439f-b979-47ea8cb6cd19 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.030380] env[62552]: DEBUG oslo_vmware.api [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for the task: (returnval){ [ 1236.030380] env[62552]: value = "task-1240247" [ 1236.030380] env[62552]: _type = "Task" [ 1236.030380] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.037864] env[62552]: DEBUG oslo_vmware.api [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.484409] env[62552]: DEBUG nova.scheduler.client.report [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1236.540444] env[62552]: DEBUG oslo_vmware.api [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Task: {'id': task-1240247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137849} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.540712] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1236.540900] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1236.541097] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1236.541274] env[62552]: INFO nova.compute.manager [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1236.541512] env[62552]: DEBUG oslo.service.loopingcall [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1236.541702] env[62552]: DEBUG nova.compute.manager [-] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1236.541797] env[62552]: DEBUG nova.network.neutron [-] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1236.838542] env[62552]: DEBUG nova.compute.manager [req-3a7edc76-17c7-46b6-a4f9-cc3c81040492 req-e2f73d0e-4231-4acf-b23e-c27bc8ea6964 service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Received event network-vif-deleted-37fd08dc-ff04-4070-858d-c78004db354d {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1236.838686] env[62552]: INFO nova.compute.manager [req-3a7edc76-17c7-46b6-a4f9-cc3c81040492 req-e2f73d0e-4231-4acf-b23e-c27bc8ea6964 service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Neutron deleted interface 37fd08dc-ff04-4070-858d-c78004db354d; detaching it from the instance and deleting it from the info cache [ 1236.838867] env[62552]: DEBUG nova.network.neutron [req-3a7edc76-17c7-46b6-a4f9-cc3c81040492 req-e2f73d0e-4231-4acf-b23e-c27bc8ea6964 service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.989522] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.990038] env[62552]: DEBUG nova.compute.manager [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1237.316783] env[62552]: DEBUG nova.network.neutron [-] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.341570] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aaad58d6-dccd-4753-beae-50d78079eb81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.352064] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49a6687-fb58-4509-9070-c7bd4d2f605c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.374991] env[62552]: DEBUG nova.compute.manager [req-3a7edc76-17c7-46b6-a4f9-cc3c81040492 req-e2f73d0e-4231-4acf-b23e-c27bc8ea6964 service nova] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Detach interface failed, port_id=37fd08dc-ff04-4070-858d-c78004db354d, reason: Instance d5d0b737-974a-4b5d-9a8e-01d101f5d476 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1237.495748] env[62552]: DEBUG nova.compute.utils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1237.496714] env[62552]: DEBUG nova.compute.manager [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1237.496913] env[62552]: DEBUG nova.network.neutron [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1237.545264] env[62552]: DEBUG nova.policy [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7bc1a611cb24ab0bf7e2e046646039c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e4d140b840b453abcd828d2129de615', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1237.799983] env[62552]: DEBUG nova.network.neutron [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Successfully created port: 152d412c-2b92-4bd8-8182-8c8761fcdbfe {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1237.819601] env[62552]: INFO nova.compute.manager [-] [instance: d5d0b737-974a-4b5d-9a8e-01d101f5d476] Took 1.28 seconds to deallocate network for instance. [ 1237.999518] env[62552]: DEBUG nova.compute.manager [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1238.326244] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.326527] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.326759] env[62552]: DEBUG nova.objects.instance [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lazy-loading 'resources' on Instance uuid d5d0b737-974a-4b5d-9a8e-01d101f5d476 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.868953] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39377a91-61e2-4fbc-9469-f4a22e47d998 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.876723] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f413dc98-ceed-4324-b893-40e18c646676 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.905869] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3718fde-7658-42ac-abc5-48890c3298c3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.912858] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522a4b7c-4207-487d-b1b0-d3098f43fa26 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.925550] env[62552]: DEBUG nova.compute.provider_tree [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1239.008985] env[62552]: DEBUG nova.compute.manager [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1239.034520] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1239.034770] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1239.034937] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1239.035141] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1239.035293] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1239.035442] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1239.035643] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1239.035804] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1239.035969] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1239.036151] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1239.036324] env[62552]: DEBUG nova.virt.hardware [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1239.037171] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b13736-a1b9-4798-a827-3963f694cb04 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.044587] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ab67d6-3bc8-465a-8e69-347f2483fd3b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.186777] env[62552]: DEBUG nova.compute.manager [req-f3c9b775-a8ab-459d-98bb-0e79c7d9b6f9 req-3bc96476-2a67-4d22-b4e4-0666446015b1 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Received event network-vif-plugged-152d412c-2b92-4bd8-8182-8c8761fcdbfe {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1239.187024] env[62552]: DEBUG oslo_concurrency.lockutils [req-f3c9b775-a8ab-459d-98bb-0e79c7d9b6f9 req-3bc96476-2a67-4d22-b4e4-0666446015b1 service nova] Acquiring lock "604e8472-7af4-49a1-9a0c-459482db9d6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.187257] env[62552]: DEBUG oslo_concurrency.lockutils [req-f3c9b775-a8ab-459d-98bb-0e79c7d9b6f9 req-3bc96476-2a67-4d22-b4e4-0666446015b1 service nova] Lock "604e8472-7af4-49a1-9a0c-459482db9d6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.187438] env[62552]: DEBUG oslo_concurrency.lockutils [req-f3c9b775-a8ab-459d-98bb-0e79c7d9b6f9 req-3bc96476-2a67-4d22-b4e4-0666446015b1 service nova] Lock "604e8472-7af4-49a1-9a0c-459482db9d6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.187612] env[62552]: DEBUG nova.compute.manager [req-f3c9b775-a8ab-459d-98bb-0e79c7d9b6f9 req-3bc96476-2a67-4d22-b4e4-0666446015b1 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] No waiting events found dispatching network-vif-plugged-152d412c-2b92-4bd8-8182-8c8761fcdbfe {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1239.187797] env[62552]: WARNING nova.compute.manager [req-f3c9b775-a8ab-459d-98bb-0e79c7d9b6f9 req-3bc96476-2a67-4d22-b4e4-0666446015b1 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Received unexpected event network-vif-plugged-152d412c-2b92-4bd8-8182-8c8761fcdbfe for instance with vm_state building and task_state spawning. [ 1239.267523] env[62552]: DEBUG nova.network.neutron [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Successfully updated port: 152d412c-2b92-4bd8-8182-8c8761fcdbfe {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1239.428304] env[62552]: DEBUG nova.scheduler.client.report [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1239.771024] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "refresh_cache-604e8472-7af4-49a1-9a0c-459482db9d6e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1239.771206] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "refresh_cache-604e8472-7af4-49a1-9a0c-459482db9d6e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.771323] env[62552]: DEBUG nova.network.neutron [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1239.933307] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.606s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.951107] env[62552]: INFO nova.scheduler.client.report [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Deleted allocations for instance d5d0b737-974a-4b5d-9a8e-01d101f5d476 [ 1240.302842] env[62552]: DEBUG nova.network.neutron [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1240.426974] env[62552]: DEBUG nova.network.neutron [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Updating instance_info_cache with network_info: [{"id": "152d412c-2b92-4bd8-8182-8c8761fcdbfe", "address": "fa:16:3e:c6:43:82", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap152d412c-2b", "ovs_interfaceid": "152d412c-2b92-4bd8-8182-8c8761fcdbfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.459216] env[62552]: DEBUG oslo_concurrency.lockutils [None req-cef749e7-c638-45fb-a2f5-612749a6f31b tempest-ServerTagsTestJSON-1318092256 tempest-ServerTagsTestJSON-1318092256-project-member] Lock "d5d0b737-974a-4b5d-9a8e-01d101f5d476" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.563s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.929714] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "refresh_cache-604e8472-7af4-49a1-9a0c-459482db9d6e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1240.930067] env[62552]: DEBUG nova.compute.manager [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Instance network_info: |[{"id": "152d412c-2b92-4bd8-8182-8c8761fcdbfe", "address": "fa:16:3e:c6:43:82", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap152d412c-2b", "ovs_interfaceid": "152d412c-2b92-4bd8-8182-8c8761fcdbfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1240.930547] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:43:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '152d412c-2b92-4bd8-8182-8c8761fcdbfe', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1240.938459] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Creating folder: Project (7e4d140b840b453abcd828d2129de615). Parent ref: group-v267339. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1240.938750] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18f73dd5-6549-43ce-8609-f03ab10f8211 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.949068] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Created folder: Project (7e4d140b840b453abcd828d2129de615) in parent group-v267339. [ 1240.949256] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Creating folder: Instances. Parent ref: group-v267541. {{(pid=62552) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1240.949490] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf82947e-9d58-4941-a67e-16dbf24f4c52 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.957627] env[62552]: INFO nova.virt.vmwareapi.vm_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Created folder: Instances in parent group-v267541. [ 1240.957854] env[62552]: DEBUG oslo.service.loopingcall [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1240.958067] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1240.958266] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e46e3e6-2c6c-4af0-8257-2b3a8af279fd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.978663] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1240.978663] env[62552]: value = "task-1240250" [ 1240.978663] env[62552]: _type = "Task" [ 1240.978663] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.986217] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240250, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.213226] env[62552]: DEBUG nova.compute.manager [req-87ff3ad1-75dd-4126-8efd-6853517ab4eb req-dd94bc2a-3596-44b4-9714-6ead28c7b8b7 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Received event network-changed-152d412c-2b92-4bd8-8182-8c8761fcdbfe {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1241.213464] env[62552]: DEBUG nova.compute.manager [req-87ff3ad1-75dd-4126-8efd-6853517ab4eb req-dd94bc2a-3596-44b4-9714-6ead28c7b8b7 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Refreshing instance network info cache due to event network-changed-152d412c-2b92-4bd8-8182-8c8761fcdbfe. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1241.213713] env[62552]: DEBUG oslo_concurrency.lockutils [req-87ff3ad1-75dd-4126-8efd-6853517ab4eb req-dd94bc2a-3596-44b4-9714-6ead28c7b8b7 service nova] Acquiring lock "refresh_cache-604e8472-7af4-49a1-9a0c-459482db9d6e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.213879] env[62552]: DEBUG oslo_concurrency.lockutils [req-87ff3ad1-75dd-4126-8efd-6853517ab4eb req-dd94bc2a-3596-44b4-9714-6ead28c7b8b7 service nova] Acquired lock "refresh_cache-604e8472-7af4-49a1-9a0c-459482db9d6e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.214081] env[62552]: DEBUG nova.network.neutron [req-87ff3ad1-75dd-4126-8efd-6853517ab4eb req-dd94bc2a-3596-44b4-9714-6ead28c7b8b7 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Refreshing network info cache for port 152d412c-2b92-4bd8-8182-8c8761fcdbfe {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1241.493290] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240250, 'name': CreateVM_Task, 'duration_secs': 0.342153} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.493640] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1241.494394] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.494652] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.495120] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1241.495459] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4c865d1-0d3d-416c-b217-1df5d65a1a75 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.500693] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1241.500693] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d1bb89-c03b-1e30-7fcb-58ef50ea5ee2" [ 1241.500693] env[62552]: _type = "Task" [ 1241.500693] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.510598] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d1bb89-c03b-1e30-7fcb-58ef50ea5ee2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.907563] env[62552]: DEBUG nova.network.neutron [req-87ff3ad1-75dd-4126-8efd-6853517ab4eb req-dd94bc2a-3596-44b4-9714-6ead28c7b8b7 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Updated VIF entry in instance network info cache for port 152d412c-2b92-4bd8-8182-8c8761fcdbfe. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1241.907928] env[62552]: DEBUG nova.network.neutron [req-87ff3ad1-75dd-4126-8efd-6853517ab4eb req-dd94bc2a-3596-44b4-9714-6ead28c7b8b7 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Updating instance_info_cache with network_info: [{"id": "152d412c-2b92-4bd8-8182-8c8761fcdbfe", "address": "fa:16:3e:c6:43:82", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap152d412c-2b", "ovs_interfaceid": "152d412c-2b92-4bd8-8182-8c8761fcdbfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.012237] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52d1bb89-c03b-1e30-7fcb-58ef50ea5ee2, 'name': SearchDatastore_Task, 'duration_secs': 0.009449} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.012237] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.012237] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1242.012446] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.012446] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.012638] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1242.012918] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04eb8527-c7e6-43f1-b5a2-9855ff3df891 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.021560] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1242.021560] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1242.022092] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ac565f7-7167-4aed-aa11-81b6b494e38f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.027750] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1242.027750] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fac3ca-172a-bfb8-bf60-d3b9d9ce900f" [ 1242.027750] env[62552]: _type = "Task" [ 1242.027750] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.035158] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fac3ca-172a-bfb8-bf60-d3b9d9ce900f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.411022] env[62552]: DEBUG oslo_concurrency.lockutils [req-87ff3ad1-75dd-4126-8efd-6853517ab4eb req-dd94bc2a-3596-44b4-9714-6ead28c7b8b7 service nova] Releasing lock "refresh_cache-604e8472-7af4-49a1-9a0c-459482db9d6e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.539234] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52fac3ca-172a-bfb8-bf60-d3b9d9ce900f, 'name': SearchDatastore_Task, 'duration_secs': 0.008803} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.540083] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1a009b4-b446-4757-a8ef-ee67c075a1c8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.545559] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1242.545559] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520ae5e1-7ec5-c66a-4675-f7bc96d71675" [ 1242.545559] env[62552]: _type = "Task" [ 1242.545559] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.553143] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520ae5e1-7ec5-c66a-4675-f7bc96d71675, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.056080] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520ae5e1-7ec5-c66a-4675-f7bc96d71675, 'name': SearchDatastore_Task, 'duration_secs': 0.008904} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.056314] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.056581] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 604e8472-7af4-49a1-9a0c-459482db9d6e/604e8472-7af4-49a1-9a0c-459482db9d6e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1243.056841] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6787fb90-4a64-41dc-9a1b-c6f07d0bffa5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.063201] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1243.063201] env[62552]: value = "task-1240251" [ 1243.063201] env[62552]: _type = "Task" [ 1243.063201] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.070344] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.576725] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240251, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438465} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.577171] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 604e8472-7af4-49a1-9a0c-459482db9d6e/604e8472-7af4-49a1-9a0c-459482db9d6e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1243.577571] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1243.577969] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d79dad6-a955-430b-9e79-b963af9a51b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.585506] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1243.585506] env[62552]: value = "task-1240252" [ 1243.585506] env[62552]: _type = "Task" [ 1243.585506] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.594785] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240252, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.094771] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240252, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062971} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.095099] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1244.095846] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e0dac8-7d03-424b-8c0b-191e86d0bfaa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.116876] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 604e8472-7af4-49a1-9a0c-459482db9d6e/604e8472-7af4-49a1-9a0c-459482db9d6e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1244.117120] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61a02aa5-b52c-43af-9e3d-dcc28bd34c81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.135748] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1244.135748] env[62552]: value = "task-1240253" [ 1244.135748] env[62552]: _type = "Task" [ 1244.135748] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.146817] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240253, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.645257] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240253, 'name': ReconfigVM_Task, 'duration_secs': 0.263702} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.645613] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 604e8472-7af4-49a1-9a0c-459482db9d6e/604e8472-7af4-49a1-9a0c-459482db9d6e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1244.646132] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6be85d67-0444-4407-afeb-9ff81ef8efe1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.652283] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1244.652283] env[62552]: value = "task-1240254" [ 1244.652283] env[62552]: _type = "Task" [ 1244.652283] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.659384] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240254, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.162074] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240254, 'name': Rename_Task, 'duration_secs': 0.147604} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.162368] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1245.162608] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57aa68e6-46e1-4a03-87b8-c0f3ddf29a4c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.169194] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1245.169194] env[62552]: value = "task-1240255" [ 1245.169194] env[62552]: _type = "Task" [ 1245.169194] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.178069] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240255, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.679317] env[62552]: DEBUG oslo_vmware.api [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240255, 'name': PowerOnVM_Task, 'duration_secs': 0.392472} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.679765] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1245.679819] env[62552]: INFO nova.compute.manager [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Took 6.67 seconds to spawn the instance on the hypervisor. [ 1245.680035] env[62552]: DEBUG nova.compute.manager [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1245.680813] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870c20c4-4fb8-4cc3-bb25-3b5ccde38979 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.196556] env[62552]: INFO nova.compute.manager [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Took 11.35 seconds to build instance. [ 1246.698321] env[62552]: DEBUG oslo_concurrency.lockutils [None req-f2c4a4c1-46ec-473d-9aed-f8462a41316c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "604e8472-7af4-49a1-9a0c-459482db9d6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.857s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.079370] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "6a6f7528-d876-4988-bb34-3bc395dd75e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.079708] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "6a6f7528-d876-4988-bb34-3bc395dd75e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1248.582799] env[62552]: DEBUG nova.compute.manager [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1249.105663] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.105949] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.107408] env[62552]: INFO nova.compute.claims [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1250.151162] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27563dce-27ce-473b-a59e-32869da8dfd7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.158534] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35c24dd-1866-4e88-9781-e8470869c751 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.188461] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18677cc9-14c8-4964-a756-646cf351f07a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.195366] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d92d8fb-7c42-4713-9b66-dce7dfae55e3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.208100] env[62552]: DEBUG nova.compute.provider_tree [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1250.711618] env[62552]: DEBUG nova.scheduler.client.report [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1251.216807] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.111s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.217371] env[62552]: DEBUG nova.compute.manager [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1251.722776] env[62552]: DEBUG nova.compute.utils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1251.723806] env[62552]: DEBUG nova.compute.manager [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1251.724394] env[62552]: DEBUG nova.network.neutron [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1251.772178] env[62552]: DEBUG nova.policy [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7bc1a611cb24ab0bf7e2e046646039c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e4d140b840b453abcd828d2129de615', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1252.006517] env[62552]: DEBUG nova.network.neutron [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Successfully created port: eb39019a-4559-4412-afd8-bf7b86e23fcb {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1252.227370] env[62552]: DEBUG nova.compute.manager [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1253.238198] env[62552]: DEBUG nova.compute.manager [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1253.263289] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1253.263507] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1253.263614] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1253.263812] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1253.263963] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1253.264128] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1253.264339] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1253.264500] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1253.264682] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1253.264870] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1253.265082] env[62552]: DEBUG nova.virt.hardware [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1253.265963] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04076afa-5838-47d8-8201-a0b185109e60 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.273851] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a622b9-7736-4bfd-a8c0-cc7b81878a83 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.369932] env[62552]: DEBUG nova.compute.manager [req-a46214a9-140c-4244-95bc-95fe4d05a512 req-faad2224-82dc-401d-a6a9-648c0a03171b service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Received event network-vif-plugged-eb39019a-4559-4412-afd8-bf7b86e23fcb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1253.370190] env[62552]: DEBUG oslo_concurrency.lockutils [req-a46214a9-140c-4244-95bc-95fe4d05a512 req-faad2224-82dc-401d-a6a9-648c0a03171b service nova] Acquiring lock "6a6f7528-d876-4988-bb34-3bc395dd75e6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.370405] env[62552]: DEBUG oslo_concurrency.lockutils [req-a46214a9-140c-4244-95bc-95fe4d05a512 req-faad2224-82dc-401d-a6a9-648c0a03171b service nova] Lock "6a6f7528-d876-4988-bb34-3bc395dd75e6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.370577] env[62552]: DEBUG oslo_concurrency.lockutils [req-a46214a9-140c-4244-95bc-95fe4d05a512 req-faad2224-82dc-401d-a6a9-648c0a03171b service nova] Lock "6a6f7528-d876-4988-bb34-3bc395dd75e6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1253.370875] env[62552]: DEBUG nova.compute.manager [req-a46214a9-140c-4244-95bc-95fe4d05a512 req-faad2224-82dc-401d-a6a9-648c0a03171b service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] No waiting events found dispatching network-vif-plugged-eb39019a-4559-4412-afd8-bf7b86e23fcb {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1253.371073] env[62552]: WARNING nova.compute.manager [req-a46214a9-140c-4244-95bc-95fe4d05a512 req-faad2224-82dc-401d-a6a9-648c0a03171b service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Received unexpected event network-vif-plugged-eb39019a-4559-4412-afd8-bf7b86e23fcb for instance with vm_state building and task_state spawning. [ 1253.445670] env[62552]: DEBUG nova.network.neutron [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Successfully updated port: eb39019a-4559-4412-afd8-bf7b86e23fcb {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1253.948379] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "refresh_cache-6a6f7528-d876-4988-bb34-3bc395dd75e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1253.948379] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "refresh_cache-6a6f7528-d876-4988-bb34-3bc395dd75e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1253.948519] env[62552]: DEBUG nova.network.neutron [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1254.479071] env[62552]: DEBUG nova.network.neutron [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1254.601206] env[62552]: DEBUG nova.network.neutron [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Updating instance_info_cache with network_info: [{"id": "eb39019a-4559-4412-afd8-bf7b86e23fcb", "address": "fa:16:3e:c9:56:08", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb39019a-45", "ovs_interfaceid": "eb39019a-4559-4412-afd8-bf7b86e23fcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.103890] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "refresh_cache-6a6f7528-d876-4988-bb34-3bc395dd75e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1255.104308] env[62552]: DEBUG nova.compute.manager [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Instance network_info: |[{"id": "eb39019a-4559-4412-afd8-bf7b86e23fcb", "address": "fa:16:3e:c9:56:08", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb39019a-45", "ovs_interfaceid": "eb39019a-4559-4412-afd8-bf7b86e23fcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1255.104756] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:56:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb39019a-4559-4412-afd8-bf7b86e23fcb', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1255.112157] env[62552]: DEBUG oslo.service.loopingcall [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1255.112376] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1255.112960] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ad8564f-d8af-40ff-92f5-3a3c79865061 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.131669] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1255.131669] env[62552]: value = "task-1240256" [ 1255.131669] env[62552]: _type = "Task" [ 1255.131669] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.138821] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240256, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.396339] env[62552]: DEBUG nova.compute.manager [req-a88e27c5-ae76-418d-a60b-5c31f27f6e62 req-36bfe1fb-7af2-4024-aa83-9b002649467d service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Received event network-changed-eb39019a-4559-4412-afd8-bf7b86e23fcb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1255.396339] env[62552]: DEBUG nova.compute.manager [req-a88e27c5-ae76-418d-a60b-5c31f27f6e62 req-36bfe1fb-7af2-4024-aa83-9b002649467d service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Refreshing instance network info cache due to event network-changed-eb39019a-4559-4412-afd8-bf7b86e23fcb. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1255.396339] env[62552]: DEBUG oslo_concurrency.lockutils [req-a88e27c5-ae76-418d-a60b-5c31f27f6e62 req-36bfe1fb-7af2-4024-aa83-9b002649467d service nova] Acquiring lock "refresh_cache-6a6f7528-d876-4988-bb34-3bc395dd75e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.396484] env[62552]: DEBUG oslo_concurrency.lockutils [req-a88e27c5-ae76-418d-a60b-5c31f27f6e62 req-36bfe1fb-7af2-4024-aa83-9b002649467d service nova] Acquired lock "refresh_cache-6a6f7528-d876-4988-bb34-3bc395dd75e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.396649] env[62552]: DEBUG nova.network.neutron [req-a88e27c5-ae76-418d-a60b-5c31f27f6e62 req-36bfe1fb-7af2-4024-aa83-9b002649467d service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Refreshing network info cache for port eb39019a-4559-4412-afd8-bf7b86e23fcb {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1255.641351] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240256, 'name': CreateVM_Task, 'duration_secs': 0.308793} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.641703] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1255.642170] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.642352] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.642683] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1255.642945] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bd97f7f-80c8-49eb-bb7a-c1997b41c124 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.647163] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1255.647163] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52605c2a-3da7-aa50-5633-d9768b6d5e44" [ 1255.647163] env[62552]: _type = "Task" [ 1255.647163] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.654284] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52605c2a-3da7-aa50-5633-d9768b6d5e44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.076591] env[62552]: DEBUG nova.network.neutron [req-a88e27c5-ae76-418d-a60b-5c31f27f6e62 req-36bfe1fb-7af2-4024-aa83-9b002649467d service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Updated VIF entry in instance network info cache for port eb39019a-4559-4412-afd8-bf7b86e23fcb. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1256.077020] env[62552]: DEBUG nova.network.neutron [req-a88e27c5-ae76-418d-a60b-5c31f27f6e62 req-36bfe1fb-7af2-4024-aa83-9b002649467d service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Updating instance_info_cache with network_info: [{"id": "eb39019a-4559-4412-afd8-bf7b86e23fcb", "address": "fa:16:3e:c9:56:08", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb39019a-45", "ovs_interfaceid": "eb39019a-4559-4412-afd8-bf7b86e23fcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.158665] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52605c2a-3da7-aa50-5633-d9768b6d5e44, 'name': SearchDatastore_Task, 'duration_secs': 0.009669} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.158981] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.159232] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1256.159547] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.159767] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.159981] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1256.160259] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed4df296-abbf-4505-b551-c5cf4db05bfb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.168071] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1256.168264] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1256.168935] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c411d6bf-4302-4d42-8531-87277bf744a0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.173579] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1256.173579] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e53703-0d6a-9877-e3f5-dfe301273003" [ 1256.173579] env[62552]: _type = "Task" [ 1256.173579] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.180774] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e53703-0d6a-9877-e3f5-dfe301273003, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.579321] env[62552]: DEBUG oslo_concurrency.lockutils [req-a88e27c5-ae76-418d-a60b-5c31f27f6e62 req-36bfe1fb-7af2-4024-aa83-9b002649467d service nova] Releasing lock "refresh_cache-6a6f7528-d876-4988-bb34-3bc395dd75e6" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.683844] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e53703-0d6a-9877-e3f5-dfe301273003, 'name': SearchDatastore_Task, 'duration_secs': 0.007559} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.684669] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d07f650-7f75-4e75-a700-e84339678dd7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.689949] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1256.689949] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a25ce0-0797-5179-38af-fab846f15f42" [ 1256.689949] env[62552]: _type = "Task" [ 1256.689949] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.697231] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a25ce0-0797-5179-38af-fab846f15f42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.199939] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52a25ce0-0797-5179-38af-fab846f15f42, 'name': SearchDatastore_Task, 'duration_secs': 0.010816} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.200306] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.200652] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6a6f7528-d876-4988-bb34-3bc395dd75e6/6a6f7528-d876-4988-bb34-3bc395dd75e6.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1257.200972] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d110a670-6723-4cee-83c1-376f00f4249b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.206767] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1257.206767] env[62552]: value = "task-1240257" [ 1257.206767] env[62552]: _type = "Task" [ 1257.206767] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.215177] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240257, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.716454] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240257, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.407988} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.716853] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 6a6f7528-d876-4988-bb34-3bc395dd75e6/6a6f7528-d876-4988-bb34-3bc395dd75e6.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1257.716927] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1257.717202] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ebcb2912-1bde-4bcd-8946-3012ba4d2344 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.723818] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1257.723818] env[62552]: value = "task-1240258" [ 1257.723818] env[62552]: _type = "Task" [ 1257.723818] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.732472] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240258, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.233842] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240258, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062102} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.234137] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1258.234897] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe30ae7-d226-4b1d-9f7d-5f9899e222e5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.257070] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 6a6f7528-d876-4988-bb34-3bc395dd75e6/6a6f7528-d876-4988-bb34-3bc395dd75e6.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1258.257319] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f5acd09-155a-44e2-8098-95569a7dc4b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.275277] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1258.275277] env[62552]: value = "task-1240259" [ 1258.275277] env[62552]: _type = "Task" [ 1258.275277] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.282474] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240259, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.785456] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240259, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.286329] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240259, 'name': ReconfigVM_Task, 'duration_secs': 0.723354} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.286602] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 6a6f7528-d876-4988-bb34-3bc395dd75e6/6a6f7528-d876-4988-bb34-3bc395dd75e6.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1259.287240] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0c09565-e8dc-408c-a482-69cf2c48b2d6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.292898] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1259.292898] env[62552]: value = "task-1240260" [ 1259.292898] env[62552]: _type = "Task" [ 1259.292898] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.300116] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240260, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.802466] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240260, 'name': Rename_Task, 'duration_secs': 0.136591} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.802839] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1259.803022] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7aad339b-cd41-4423-bd42-19ff883e5134 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.808536] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1259.808536] env[62552]: value = "task-1240261" [ 1259.808536] env[62552]: _type = "Task" [ 1259.808536] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.816669] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240261, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.320334] env[62552]: DEBUG oslo_vmware.api [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240261, 'name': PowerOnVM_Task, 'duration_secs': 0.400755} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.320661] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1260.320918] env[62552]: INFO nova.compute.manager [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Took 7.08 seconds to spawn the instance on the hypervisor. [ 1260.321221] env[62552]: DEBUG nova.compute.manager [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1260.322068] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0f0bbc-526a-4182-9d3f-594dbb578683 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.842584] env[62552]: INFO nova.compute.manager [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Took 11.75 seconds to build instance. [ 1261.341535] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "6a6f7528-d876-4988-bb34-3bc395dd75e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.344188] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eb0b5225-de69-4ec4-b5f5-0303e8b42264 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "6a6f7528-d876-4988-bb34-3bc395dd75e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.264s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.344435] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "6a6f7528-d876-4988-bb34-3bc395dd75e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.344655] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "6a6f7528-d876-4988-bb34-3bc395dd75e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.344860] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "6a6f7528-d876-4988-bb34-3bc395dd75e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.345040] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "6a6f7528-d876-4988-bb34-3bc395dd75e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.347050] env[62552]: INFO nova.compute.manager [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Terminating instance [ 1261.850449] env[62552]: DEBUG nova.compute.manager [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1261.850836] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1261.851625] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5eca05-d07b-4fdd-87b0-37b62f9084a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.859711] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1261.859955] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67b7f77a-632c-4071-b5ff-90fe9948636a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.866145] env[62552]: DEBUG oslo_vmware.api [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1261.866145] env[62552]: value = "task-1240262" [ 1261.866145] env[62552]: _type = "Task" [ 1261.866145] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.874333] env[62552]: DEBUG oslo_vmware.api [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240262, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.375468] env[62552]: DEBUG oslo_vmware.api [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240262, 'name': PowerOffVM_Task, 'duration_secs': 0.202477} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.375736] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1262.375905] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1262.376163] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3a8c3a3-bd7d-4f91-8e16-06b29bdfa61b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.438924] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1262.439126] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1262.439274] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleting the datastore file [datastore2] 6a6f7528-d876-4988-bb34-3bc395dd75e6 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1262.439538] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f28174d-690c-4e08-9f89-b25037da6fef {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.445762] env[62552]: DEBUG oslo_vmware.api [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1262.445762] env[62552]: value = "task-1240264" [ 1262.445762] env[62552]: _type = "Task" [ 1262.445762] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.453316] env[62552]: DEBUG oslo_vmware.api [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240264, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.955772] env[62552]: DEBUG oslo_vmware.api [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240264, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142084} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.956313] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1262.956590] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1262.956782] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1262.956963] env[62552]: INFO nova.compute.manager [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1262.957225] env[62552]: DEBUG oslo.service.loopingcall [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1262.957501] env[62552]: DEBUG nova.compute.manager [-] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1262.957621] env[62552]: DEBUG nova.network.neutron [-] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1263.213985] env[62552]: DEBUG nova.compute.manager [req-03d2ca7f-71bd-4f09-b517-42d89d7bf28f req-d66fd90f-4929-40c4-97f0-8ac64e15f78f service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Received event network-vif-deleted-eb39019a-4559-4412-afd8-bf7b86e23fcb {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1263.214069] env[62552]: INFO nova.compute.manager [req-03d2ca7f-71bd-4f09-b517-42d89d7bf28f req-d66fd90f-4929-40c4-97f0-8ac64e15f78f service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Neutron deleted interface eb39019a-4559-4412-afd8-bf7b86e23fcb; detaching it from the instance and deleting it from the info cache [ 1263.214252] env[62552]: DEBUG nova.network.neutron [req-03d2ca7f-71bd-4f09-b517-42d89d7bf28f req-d66fd90f-4929-40c4-97f0-8ac64e15f78f service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.694182] env[62552]: DEBUG nova.network.neutron [-] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.716872] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac0c12a3-0916-4949-9180-a4965f9de389 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.725753] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b80e1bc-b425-46d2-b3ab-501313a83158 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.747809] env[62552]: DEBUG nova.compute.manager [req-03d2ca7f-71bd-4f09-b517-42d89d7bf28f req-d66fd90f-4929-40c4-97f0-8ac64e15f78f service nova] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Detach interface failed, port_id=eb39019a-4559-4412-afd8-bf7b86e23fcb, reason: Instance 6a6f7528-d876-4988-bb34-3bc395dd75e6 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1264.197059] env[62552]: INFO nova.compute.manager [-] [instance: 6a6f7528-d876-4988-bb34-3bc395dd75e6] Took 1.24 seconds to deallocate network for instance. [ 1264.705050] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.705050] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.705050] env[62552]: DEBUG nova.objects.instance [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lazy-loading 'resources' on Instance uuid 6a6f7528-d876-4988-bb34-3bc395dd75e6 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1265.255940] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f860ac-4952-4cf9-b135-113b38b43f4c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.262973] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecda3faa-0acd-4389-b0ed-7b07b7b3a3a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.292321] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddf6187-8bf2-4122-a5fa-a21ae3de449a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.299241] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f8f06a-91d6-4682-a68e-f418ab57ec13 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.313655] env[62552]: DEBUG nova.compute.provider_tree [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1265.816648] env[62552]: DEBUG nova.scheduler.client.report [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1266.321721] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.342158] env[62552]: INFO nova.scheduler.client.report [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted allocations for instance 6a6f7528-d876-4988-bb34-3bc395dd75e6 [ 1266.850408] env[62552]: DEBUG oslo_concurrency.lockutils [None req-6d3eec5f-861f-4b55-aba2-da970d3c17ec tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "6a6f7528-d876-4988-bb34-3bc395dd75e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.506s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.332891] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.333176] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.835694] env[62552]: DEBUG nova.compute.manager [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1269.358049] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.358332] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.359869] env[62552]: INFO nova.compute.claims [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1270.407582] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2c2810-1c4c-4def-a90e-394c124b6cd3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.416466] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a26599a-ef7f-4133-b9ed-8902c742102f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.449434] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf309d79-fb75-4fd3-97ed-c0d0be443e45 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.456620] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151709aa-e808-4a59-9a91-2f89efc56e6b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.469474] env[62552]: DEBUG nova.compute.provider_tree [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.972862] env[62552]: DEBUG nova.scheduler.client.report [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1271.477672] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.478216] env[62552]: DEBUG nova.compute.manager [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1271.983131] env[62552]: DEBUG nova.compute.utils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1271.988019] env[62552]: DEBUG nova.compute.manager [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1271.988019] env[62552]: DEBUG nova.network.neutron [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1272.034324] env[62552]: DEBUG nova.policy [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7bc1a611cb24ab0bf7e2e046646039c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e4d140b840b453abcd828d2129de615', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1272.271331] env[62552]: DEBUG nova.network.neutron [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Successfully created port: b2b9c122-0335-4ece-8b08-9fb15fec626f {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1272.488394] env[62552]: DEBUG nova.compute.manager [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1273.498788] env[62552]: DEBUG nova.compute.manager [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1273.525324] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1273.525576] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1273.525742] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1273.525933] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1273.526124] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1273.526329] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1273.526492] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1273.526659] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1273.526828] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1273.526994] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1273.527192] env[62552]: DEBUG nova.virt.hardware [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1273.528871] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffad783-8412-41b7-a4a2-eff0fc96027b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.536661] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb5ead3-7d95-440a-9f03-27597f8cb6f9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.629211] env[62552]: DEBUG nova.compute.manager [req-18d3e3eb-bf71-49ae-86fe-7ca7bac9db5a req-33caa484-2789-48a0-bac9-15b9347716d6 service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Received event network-vif-plugged-b2b9c122-0335-4ece-8b08-9fb15fec626f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1273.629444] env[62552]: DEBUG oslo_concurrency.lockutils [req-18d3e3eb-bf71-49ae-86fe-7ca7bac9db5a req-33caa484-2789-48a0-bac9-15b9347716d6 service nova] Acquiring lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1273.629697] env[62552]: DEBUG oslo_concurrency.lockutils [req-18d3e3eb-bf71-49ae-86fe-7ca7bac9db5a req-33caa484-2789-48a0-bac9-15b9347716d6 service nova] Lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.629925] env[62552]: DEBUG oslo_concurrency.lockutils [req-18d3e3eb-bf71-49ae-86fe-7ca7bac9db5a req-33caa484-2789-48a0-bac9-15b9347716d6 service nova] Lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.630121] env[62552]: DEBUG nova.compute.manager [req-18d3e3eb-bf71-49ae-86fe-7ca7bac9db5a req-33caa484-2789-48a0-bac9-15b9347716d6 service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] No waiting events found dispatching network-vif-plugged-b2b9c122-0335-4ece-8b08-9fb15fec626f {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1273.630299] env[62552]: WARNING nova.compute.manager [req-18d3e3eb-bf71-49ae-86fe-7ca7bac9db5a req-33caa484-2789-48a0-bac9-15b9347716d6 service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Received unexpected event network-vif-plugged-b2b9c122-0335-4ece-8b08-9fb15fec626f for instance with vm_state building and task_state spawning. [ 1273.710959] env[62552]: DEBUG nova.network.neutron [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Successfully updated port: b2b9c122-0335-4ece-8b08-9fb15fec626f {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1274.213931] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "refresh_cache-54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1274.213931] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "refresh_cache-54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.214119] env[62552]: DEBUG nova.network.neutron [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1274.744330] env[62552]: DEBUG nova.network.neutron [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1274.860776] env[62552]: DEBUG nova.network.neutron [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Updating instance_info_cache with network_info: [{"id": "b2b9c122-0335-4ece-8b08-9fb15fec626f", "address": "fa:16:3e:8a:d7:c1", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2b9c122-03", "ovs_interfaceid": "b2b9c122-0335-4ece-8b08-9fb15fec626f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1275.363496] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "refresh_cache-54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1275.363823] env[62552]: DEBUG nova.compute.manager [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Instance network_info: |[{"id": "b2b9c122-0335-4ece-8b08-9fb15fec626f", "address": "fa:16:3e:8a:d7:c1", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2b9c122-03", "ovs_interfaceid": "b2b9c122-0335-4ece-8b08-9fb15fec626f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1275.364289] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:d7:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2b9c122-0335-4ece-8b08-9fb15fec626f', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1275.371819] env[62552]: DEBUG oslo.service.loopingcall [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1275.372040] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1275.372279] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68d70302-100d-4f49-8346-f58d3d558e17 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.391966] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1275.391966] env[62552]: value = "task-1240265" [ 1275.391966] env[62552]: _type = "Task" [ 1275.391966] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.399144] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240265, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.653305] env[62552]: DEBUG nova.compute.manager [req-4d997e6b-9c54-463a-a20f-c80eab1716d2 req-c76ef261-9dc2-421c-a006-dd6815091dbb service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Received event network-changed-b2b9c122-0335-4ece-8b08-9fb15fec626f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1275.653543] env[62552]: DEBUG nova.compute.manager [req-4d997e6b-9c54-463a-a20f-c80eab1716d2 req-c76ef261-9dc2-421c-a006-dd6815091dbb service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Refreshing instance network info cache due to event network-changed-b2b9c122-0335-4ece-8b08-9fb15fec626f. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1275.653787] env[62552]: DEBUG oslo_concurrency.lockutils [req-4d997e6b-9c54-463a-a20f-c80eab1716d2 req-c76ef261-9dc2-421c-a006-dd6815091dbb service nova] Acquiring lock "refresh_cache-54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1275.653940] env[62552]: DEBUG oslo_concurrency.lockutils [req-4d997e6b-9c54-463a-a20f-c80eab1716d2 req-c76ef261-9dc2-421c-a006-dd6815091dbb service nova] Acquired lock "refresh_cache-54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.654131] env[62552]: DEBUG nova.network.neutron [req-4d997e6b-9c54-463a-a20f-c80eab1716d2 req-c76ef261-9dc2-421c-a006-dd6815091dbb service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Refreshing network info cache for port b2b9c122-0335-4ece-8b08-9fb15fec626f {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1275.902158] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240265, 'name': CreateVM_Task, 'duration_secs': 0.310411} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.902480] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1275.902967] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1275.903156] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.903530] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1275.903783] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f66b77f-0e79-4e42-814a-ab962dbc2f03 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.908125] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1275.908125] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cb9d00-5c62-3dbb-8b6f-2355d6642aeb" [ 1275.908125] env[62552]: _type = "Task" [ 1275.908125] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.915520] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cb9d00-5c62-3dbb-8b6f-2355d6642aeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.331180] env[62552]: DEBUG nova.network.neutron [req-4d997e6b-9c54-463a-a20f-c80eab1716d2 req-c76ef261-9dc2-421c-a006-dd6815091dbb service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Updated VIF entry in instance network info cache for port b2b9c122-0335-4ece-8b08-9fb15fec626f. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1276.331546] env[62552]: DEBUG nova.network.neutron [req-4d997e6b-9c54-463a-a20f-c80eab1716d2 req-c76ef261-9dc2-421c-a006-dd6815091dbb service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Updating instance_info_cache with network_info: [{"id": "b2b9c122-0335-4ece-8b08-9fb15fec626f", "address": "fa:16:3e:8a:d7:c1", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2b9c122-03", "ovs_interfaceid": "b2b9c122-0335-4ece-8b08-9fb15fec626f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1276.417485] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52cb9d00-5c62-3dbb-8b6f-2355d6642aeb, 'name': SearchDatastore_Task, 'duration_secs': 0.009361} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.417847] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1276.418105] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1276.418392] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1276.418495] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1276.418684] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1276.418946] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce7c24f9-f564-4dde-85f1-d8e9bce597d3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.426750] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1276.426934] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1276.427653] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a82e1ec0-f50b-4531-9a0b-29e308007749 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.432169] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1276.432169] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520bc8f1-bf50-9732-26b9-a63c1dd4d14b" [ 1276.432169] env[62552]: _type = "Task" [ 1276.432169] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.440348] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520bc8f1-bf50-9732-26b9-a63c1dd4d14b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.834878] env[62552]: DEBUG oslo_concurrency.lockutils [req-4d997e6b-9c54-463a-a20f-c80eab1716d2 req-c76ef261-9dc2-421c-a006-dd6815091dbb service nova] Releasing lock "refresh_cache-54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1276.942409] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520bc8f1-bf50-9732-26b9-a63c1dd4d14b, 'name': SearchDatastore_Task, 'duration_secs': 0.008629} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.943175] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6305b7f3-0639-41ea-b1d4-c2b6b911d10b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.947825] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1276.947825] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c5deaf-2206-20bc-ee45-2c78659d5c74" [ 1276.947825] env[62552]: _type = "Task" [ 1276.947825] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.954587] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c5deaf-2206-20bc-ee45-2c78659d5c74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.458612] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c5deaf-2206-20bc-ee45-2c78659d5c74, 'name': SearchDatastore_Task, 'duration_secs': 0.008436} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.458874] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1277.459150] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3/54a57896-f8cf-4dbd-8f1d-8bb7509d14d3.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1277.459409] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1fe1223-b1a4-4a1d-86f7-6ad6ddf48f5c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.465917] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1277.465917] env[62552]: value = "task-1240266" [ 1277.465917] env[62552]: _type = "Task" [ 1277.465917] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.473187] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240266, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.975461] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240266, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.422423} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.975828] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3/54a57896-f8cf-4dbd-8f1d-8bb7509d14d3.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1277.975955] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1277.976221] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18d31ae9-0c1d-4a36-b576-8be725040209 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.981894] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1277.981894] env[62552]: value = "task-1240267" [ 1277.981894] env[62552]: _type = "Task" [ 1277.981894] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.988543] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240267, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.491796] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240267, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086534} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.492065] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1278.492795] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a533b20c-ce55-4a31-9c15-dd2cceb488cf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.514018] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3/54a57896-f8cf-4dbd-8f1d-8bb7509d14d3.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1278.514264] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b27b8fee-d098-4d57-a511-f7ccf63ed10b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.532974] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1278.532974] env[62552]: value = "task-1240268" [ 1278.532974] env[62552]: _type = "Task" [ 1278.532974] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.539991] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240268, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.042807] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240268, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.542999] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240268, 'name': ReconfigVM_Task, 'duration_secs': 0.666131} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.543343] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3/54a57896-f8cf-4dbd-8f1d-8bb7509d14d3.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1279.543979] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd0928e7-9ef6-434e-9527-daf1962530fa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.550664] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1279.550664] env[62552]: value = "task-1240269" [ 1279.550664] env[62552]: _type = "Task" [ 1279.550664] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.558290] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240269, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.698080] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.698341] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.060216] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240269, 'name': Rename_Task, 'duration_secs': 0.134075} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.060569] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1280.060724] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-264b8638-e1ff-414e-b4f3-81eaa0a98cd8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.066278] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1280.066278] env[62552]: value = "task-1240270" [ 1280.066278] env[62552]: _type = "Task" [ 1280.066278] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.074149] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240270, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.577049] env[62552]: DEBUG oslo_vmware.api [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240270, 'name': PowerOnVM_Task, 'duration_secs': 0.442755} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.577325] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1280.577530] env[62552]: INFO nova.compute.manager [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Took 7.08 seconds to spawn the instance on the hypervisor. [ 1280.577713] env[62552]: DEBUG nova.compute.manager [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1280.578471] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bcf7f0-dbc3-43f0-a1b2-897d394836da {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.698122] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.698290] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1280.698381] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Rebuilding the list of instances to heal {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1281.097291] env[62552]: INFO nova.compute.manager [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Took 11.76 seconds to build instance. [ 1281.226794] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "refresh_cache-604e8472-7af4-49a1-9a0c-459482db9d6e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1281.226946] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquired lock "refresh_cache-604e8472-7af4-49a1-9a0c-459482db9d6e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.227106] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Forcefully refreshing network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1281.227279] env[62552]: DEBUG nova.objects.instance [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lazy-loading 'info_cache' on Instance uuid 604e8472-7af4-49a1-9a0c-459482db9d6e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1281.599552] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9d833ebf-a7a2-4453-9a6b-b792d0e471c2 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.266s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.763506] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.763771] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.763984] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.764188] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.764363] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.766320] env[62552]: INFO nova.compute.manager [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Terminating instance [ 1282.269252] env[62552]: DEBUG nova.compute.manager [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1282.269565] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1282.270354] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f24820a-a1f7-455b-ac57-ee0ffcbb647c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.279217] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1282.279457] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-123c25a0-a2d0-4d45-939d-2c4ac06cde56 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.286743] env[62552]: DEBUG oslo_vmware.api [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1282.286743] env[62552]: value = "task-1240271" [ 1282.286743] env[62552]: _type = "Task" [ 1282.286743] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.294944] env[62552]: DEBUG oslo_vmware.api [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.796289] env[62552]: DEBUG oslo_vmware.api [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240271, 'name': PowerOffVM_Task, 'duration_secs': 0.172607} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.796580] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1282.796758] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1282.797017] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9ca9cba-9459-422e-8a79-92c2cb0eff87 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.856231] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1282.856452] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1282.856656] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleting the datastore file [datastore2] 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1282.856918] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58c34260-0385-45dc-a391-9fa84fbf5d8a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.862749] env[62552]: DEBUG oslo_vmware.api [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1282.862749] env[62552]: value = "task-1240273" [ 1282.862749] env[62552]: _type = "Task" [ 1282.862749] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.871038] env[62552]: DEBUG oslo_vmware.api [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240273, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.935346] env[62552]: DEBUG nova.network.neutron [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Updating instance_info_cache with network_info: [{"id": "152d412c-2b92-4bd8-8182-8c8761fcdbfe", "address": "fa:16:3e:c6:43:82", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap152d412c-2b", "ovs_interfaceid": "152d412c-2b92-4bd8-8182-8c8761fcdbfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1283.372190] env[62552]: DEBUG oslo_vmware.api [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240273, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1187} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.372561] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1283.372619] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1283.372790] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1283.372967] env[62552]: INFO nova.compute.manager [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1283.373232] env[62552]: DEBUG oslo.service.loopingcall [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1283.373460] env[62552]: DEBUG nova.compute.manager [-] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1283.373562] env[62552]: DEBUG nova.network.neutron [-] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1283.437705] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Releasing lock "refresh_cache-604e8472-7af4-49a1-9a0c-459482db9d6e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1283.437915] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Updated the network info_cache for instance {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1283.438125] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.438285] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.438445] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.438572] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1283.438716] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.637314] env[62552]: DEBUG nova.compute.manager [req-c5cf5aa3-15c3-4556-9fd9-a4610104c875 req-303a6927-8dcb-4b0f-b3ef-d542620558ca service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Received event network-vif-deleted-b2b9c122-0335-4ece-8b08-9fb15fec626f {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1283.637518] env[62552]: INFO nova.compute.manager [req-c5cf5aa3-15c3-4556-9fd9-a4610104c875 req-303a6927-8dcb-4b0f-b3ef-d542620558ca service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Neutron deleted interface b2b9c122-0335-4ece-8b08-9fb15fec626f; detaching it from the instance and deleting it from the info cache [ 1283.638066] env[62552]: DEBUG nova.network.neutron [req-c5cf5aa3-15c3-4556-9fd9-a4610104c875 req-303a6927-8dcb-4b0f-b3ef-d542620558ca service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1283.941814] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1283.941990] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1283.942181] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1283.942339] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1283.943258] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3b2a72-011b-419e-b068-2a40d9db9799 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.951488] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de2adc92-e454-4afe-8d64-a2b2733c621c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.965203] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b417d6-c631-4a3b-9dbe-665792adc0e7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.971693] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9f218c-f781-4965-aa79-cdef7455124f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.000731] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181281MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1284.000916] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1284.001125] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1284.114160] env[62552]: DEBUG nova.network.neutron [-] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.139913] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9fa01037-eca5-4d34-bbcc-f404dd6c38f6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.149638] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d155a6-502c-4392-be09-72c15648ec48 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.173529] env[62552]: DEBUG nova.compute.manager [req-c5cf5aa3-15c3-4556-9fd9-a4610104c875 req-303a6927-8dcb-4b0f-b3ef-d542620558ca service nova] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Detach interface failed, port_id=b2b9c122-0335-4ece-8b08-9fb15fec626f, reason: Instance 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1284.618063] env[62552]: INFO nova.compute.manager [-] [instance: 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3] Took 1.24 seconds to deallocate network for instance. [ 1285.024548] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 604e8472-7af4-49a1-9a0c-459482db9d6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1285.024716] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1285.024899] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1285.025472] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1285.058789] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e250267f-92e2-40ab-aaa0-8b4197fb1058 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.066017] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de24433-da9a-4e4d-8289-05b6e16310bf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.094543] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735dff6f-48f3-444a-a52a-6579ec7f137a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.100953] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed43a33a-9cc0-495b-9e0d-f10f0b203282 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.113119] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1285.122389] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1285.616732] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1286.121386] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1286.121844] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.120s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1286.121899] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1286.122110] env[62552]: DEBUG nova.objects.instance [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lazy-loading 'resources' on Instance uuid 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1286.659209] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b69a19-b06e-46ce-8575-d7c3beb7b428 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.667056] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99dce9d7-042b-4895-86c7-cfe719ed5005 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.695731] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e988b0-1029-433d-b701-353fd68ea1e2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.702147] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d702a4da-7a23-415f-90f4-245d55a3daf2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.714347] env[62552]: DEBUG nova.compute.provider_tree [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1287.118563] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1287.217199] env[62552]: DEBUG nova.scheduler.client.report [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1287.697713] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1287.721783] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1287.739779] env[62552]: INFO nova.scheduler.client.report [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted allocations for instance 54a57896-f8cf-4dbd-8f1d-8bb7509d14d3 [ 1288.249089] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d473733a-08f4-4b49-b62c-fac86105ae95 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "54a57896-f8cf-4dbd-8f1d-8bb7509d14d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.485s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1288.760703] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "a45dba31-8b95-458e-a40d-dae81b26c27a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1288.760937] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "a45dba31-8b95-458e-a40d-dae81b26c27a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1289.263820] env[62552]: DEBUG nova.compute.manager [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1289.782632] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1289.782904] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1289.784359] env[62552]: INFO nova.compute.claims [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1290.826232] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5d0e3c-c792-426e-bd26-88c03beb0dd0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.833838] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a844911-958b-4964-a925-7de236d8dc20 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.862138] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b89fc67-be2d-47c5-8b18-b6d76b487c7d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.868909] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d60b7a6-90bc-4ace-a04c-4fd93ea84c04 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.883091] env[62552]: DEBUG nova.compute.provider_tree [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1291.386445] env[62552]: DEBUG nova.scheduler.client.report [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1291.891430] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.108s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1291.891937] env[62552]: DEBUG nova.compute.manager [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1292.396549] env[62552]: DEBUG nova.compute.utils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1292.397943] env[62552]: DEBUG nova.compute.manager [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1292.398121] env[62552]: DEBUG nova.network.neutron [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1292.450420] env[62552]: DEBUG nova.policy [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7bc1a611cb24ab0bf7e2e046646039c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e4d140b840b453abcd828d2129de615', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1292.689321] env[62552]: DEBUG nova.network.neutron [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Successfully created port: 5a32c6a9-da6a-46be-9196-ac1d7554ad20 {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1292.901823] env[62552]: DEBUG nova.compute.manager [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1293.955726] env[62552]: DEBUG nova.compute.manager [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1293.955726] env[62552]: DEBUG nova.virt.hardware [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1293.955726] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdd245d-4949-40ac-a6be-5c900076ed14 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.955726] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433ada34-1d40-4c6d-90b1-3b8b348d6b79 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.048176] env[62552]: DEBUG nova.compute.manager [req-15e18998-3763-436e-9132-3f1c0ae480b4 req-479e7bd4-1fe6-4cc4-ba0f-a2ea17a2d74d service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Received event network-vif-plugged-5a32c6a9-da6a-46be-9196-ac1d7554ad20 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1294.048418] env[62552]: DEBUG oslo_concurrency.lockutils [req-15e18998-3763-436e-9132-3f1c0ae480b4 req-479e7bd4-1fe6-4cc4-ba0f-a2ea17a2d74d service nova] Acquiring lock "a45dba31-8b95-458e-a40d-dae81b26c27a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1294.048632] env[62552]: DEBUG oslo_concurrency.lockutils [req-15e18998-3763-436e-9132-3f1c0ae480b4 req-479e7bd4-1fe6-4cc4-ba0f-a2ea17a2d74d service nova] Lock "a45dba31-8b95-458e-a40d-dae81b26c27a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1294.048852] env[62552]: DEBUG oslo_concurrency.lockutils [req-15e18998-3763-436e-9132-3f1c0ae480b4 req-479e7bd4-1fe6-4cc4-ba0f-a2ea17a2d74d service nova] Lock "a45dba31-8b95-458e-a40d-dae81b26c27a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1294.049122] env[62552]: DEBUG nova.compute.manager [req-15e18998-3763-436e-9132-3f1c0ae480b4 req-479e7bd4-1fe6-4cc4-ba0f-a2ea17a2d74d service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] No waiting events found dispatching network-vif-plugged-5a32c6a9-da6a-46be-9196-ac1d7554ad20 {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1294.049217] env[62552]: WARNING nova.compute.manager [req-15e18998-3763-436e-9132-3f1c0ae480b4 req-479e7bd4-1fe6-4cc4-ba0f-a2ea17a2d74d service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Received unexpected event network-vif-plugged-5a32c6a9-da6a-46be-9196-ac1d7554ad20 for instance with vm_state building and task_state spawning. [ 1294.132997] env[62552]: DEBUG nova.network.neutron [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Successfully updated port: 5a32c6a9-da6a-46be-9196-ac1d7554ad20 {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1294.635778] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "refresh_cache-a45dba31-8b95-458e-a40d-dae81b26c27a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1294.635827] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "refresh_cache-a45dba31-8b95-458e-a40d-dae81b26c27a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1294.635967] env[62552]: DEBUG nova.network.neutron [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1295.167646] env[62552]: DEBUG nova.network.neutron [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1295.283717] env[62552]: DEBUG nova.network.neutron [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Updating instance_info_cache with network_info: [{"id": "5a32c6a9-da6a-46be-9196-ac1d7554ad20", "address": "fa:16:3e:12:ba:f8", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a32c6a9-da", "ovs_interfaceid": "5a32c6a9-da6a-46be-9196-ac1d7554ad20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1295.786562] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "refresh_cache-a45dba31-8b95-458e-a40d-dae81b26c27a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1295.786869] env[62552]: DEBUG nova.compute.manager [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Instance network_info: |[{"id": "5a32c6a9-da6a-46be-9196-ac1d7554ad20", "address": "fa:16:3e:12:ba:f8", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a32c6a9-da", "ovs_interfaceid": "5a32c6a9-da6a-46be-9196-ac1d7554ad20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1295.787319] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:ba:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a32c6a9-da6a-46be-9196-ac1d7554ad20', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1295.794650] env[62552]: DEBUG oslo.service.loopingcall [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1295.794858] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1295.795094] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe4fea2d-f0aa-47ca-aff3-c7e71b5abaa2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.814398] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1295.814398] env[62552]: value = "task-1240274" [ 1295.814398] env[62552]: _type = "Task" [ 1295.814398] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.821469] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240274, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.075809] env[62552]: DEBUG nova.compute.manager [req-b86baa90-f676-4b4d-88cf-76e7f7d012a1 req-78f0279e-e413-45d2-b77b-95dde30fee29 service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Received event network-changed-5a32c6a9-da6a-46be-9196-ac1d7554ad20 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1296.076036] env[62552]: DEBUG nova.compute.manager [req-b86baa90-f676-4b4d-88cf-76e7f7d012a1 req-78f0279e-e413-45d2-b77b-95dde30fee29 service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Refreshing instance network info cache due to event network-changed-5a32c6a9-da6a-46be-9196-ac1d7554ad20. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1296.076262] env[62552]: DEBUG oslo_concurrency.lockutils [req-b86baa90-f676-4b4d-88cf-76e7f7d012a1 req-78f0279e-e413-45d2-b77b-95dde30fee29 service nova] Acquiring lock "refresh_cache-a45dba31-8b95-458e-a40d-dae81b26c27a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1296.076412] env[62552]: DEBUG oslo_concurrency.lockutils [req-b86baa90-f676-4b4d-88cf-76e7f7d012a1 req-78f0279e-e413-45d2-b77b-95dde30fee29 service nova] Acquired lock "refresh_cache-a45dba31-8b95-458e-a40d-dae81b26c27a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1296.076579] env[62552]: DEBUG nova.network.neutron [req-b86baa90-f676-4b4d-88cf-76e7f7d012a1 req-78f0279e-e413-45d2-b77b-95dde30fee29 service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Refreshing network info cache for port 5a32c6a9-da6a-46be-9196-ac1d7554ad20 {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1296.324193] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240274, 'name': CreateVM_Task, 'duration_secs': 0.288049} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.324526] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1296.324996] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1296.325193] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1296.325544] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1296.325802] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da3df85c-3e78-4c33-a37a-bb9329e3668e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.329913] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1296.329913] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52811718-3114-69aa-3542-86c4c4e6b14a" [ 1296.329913] env[62552]: _type = "Task" [ 1296.329913] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.337402] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52811718-3114-69aa-3542-86c4c4e6b14a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.752317] env[62552]: DEBUG nova.network.neutron [req-b86baa90-f676-4b4d-88cf-76e7f7d012a1 req-78f0279e-e413-45d2-b77b-95dde30fee29 service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Updated VIF entry in instance network info cache for port 5a32c6a9-da6a-46be-9196-ac1d7554ad20. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1296.752680] env[62552]: DEBUG nova.network.neutron [req-b86baa90-f676-4b4d-88cf-76e7f7d012a1 req-78f0279e-e413-45d2-b77b-95dde30fee29 service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Updating instance_info_cache with network_info: [{"id": "5a32c6a9-da6a-46be-9196-ac1d7554ad20", "address": "fa:16:3e:12:ba:f8", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a32c6a9-da", "ovs_interfaceid": "5a32c6a9-da6a-46be-9196-ac1d7554ad20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1296.839848] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52811718-3114-69aa-3542-86c4c4e6b14a, 'name': SearchDatastore_Task, 'duration_secs': 0.008723} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.840161] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1296.840401] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1296.840635] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1296.840790] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1296.840970] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1296.841240] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6ba705a-9288-4320-82f8-1ef41863bbd3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.849654] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1296.849833] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1296.850534] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d83d7643-4d9f-4c43-b3dd-54e074011e65 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.855901] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1296.855901] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bf2eca-7819-e63b-6cf1-f28dba011752" [ 1296.855901] env[62552]: _type = "Task" [ 1296.855901] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.863966] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bf2eca-7819-e63b-6cf1-f28dba011752, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.255843] env[62552]: DEBUG oslo_concurrency.lockutils [req-b86baa90-f676-4b4d-88cf-76e7f7d012a1 req-78f0279e-e413-45d2-b77b-95dde30fee29 service nova] Releasing lock "refresh_cache-a45dba31-8b95-458e-a40d-dae81b26c27a" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1297.366973] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52bf2eca-7819-e63b-6cf1-f28dba011752, 'name': SearchDatastore_Task, 'duration_secs': 0.008589} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1297.367858] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57bb8bde-5e4d-4ba4-ae2b-e1c6548207a6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.373218] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1297.373218] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5294eee6-8231-4ca6-9f7e-44420a27b9db" [ 1297.373218] env[62552]: _type = "Task" [ 1297.373218] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.381394] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5294eee6-8231-4ca6-9f7e-44420a27b9db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.884152] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5294eee6-8231-4ca6-9f7e-44420a27b9db, 'name': SearchDatastore_Task, 'duration_secs': 0.009186} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1297.884440] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1297.884703] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] a45dba31-8b95-458e-a40d-dae81b26c27a/a45dba31-8b95-458e-a40d-dae81b26c27a.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1297.884975] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f002ecc5-62f4-44a7-837e-0952d51406ab {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.891545] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1297.891545] env[62552]: value = "task-1240275" [ 1297.891545] env[62552]: _type = "Task" [ 1297.891545] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.899268] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240275, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.401471] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240275, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443926} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.401854] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] a45dba31-8b95-458e-a40d-dae81b26c27a/a45dba31-8b95-458e-a40d-dae81b26c27a.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1298.401999] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1298.402313] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4b5ac73-d8dc-4999-bd52-925a75ceea9e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.409839] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1298.409839] env[62552]: value = "task-1240276" [ 1298.409839] env[62552]: _type = "Task" [ 1298.409839] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.417207] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240276, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.919244] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240276, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061949} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.919514] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1298.920242] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1c731f-a51d-4df5-acc0-f7921fd7d083 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.940647] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] a45dba31-8b95-458e-a40d-dae81b26c27a/a45dba31-8b95-458e-a40d-dae81b26c27a.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1298.940863] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-463967c0-2459-42fa-a550-39d2b4e3018e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.959398] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1298.959398] env[62552]: value = "task-1240277" [ 1298.959398] env[62552]: _type = "Task" [ 1298.959398] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.966259] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240277, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.469152] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240277, 'name': ReconfigVM_Task, 'duration_secs': 0.313991} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.469552] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Reconfigured VM instance instance-00000072 to attach disk [datastore2] a45dba31-8b95-458e-a40d-dae81b26c27a/a45dba31-8b95-458e-a40d-dae81b26c27a.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1299.470095] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b601c45e-c278-4765-a730-716ef8bde0b7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.475922] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1299.475922] env[62552]: value = "task-1240278" [ 1299.475922] env[62552]: _type = "Task" [ 1299.475922] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1299.483816] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240278, 'name': Rename_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.985671] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240278, 'name': Rename_Task, 'duration_secs': 0.129343} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.985966] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1299.986224] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36d90db7-53a3-44ad-8776-3ad82a5668ad {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.991542] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1299.991542] env[62552]: value = "task-1240279" [ 1299.991542] env[62552]: _type = "Task" [ 1299.991542] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1299.998527] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240279, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.501226] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240279, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.004882] env[62552]: DEBUG oslo_vmware.api [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240279, 'name': PowerOnVM_Task, 'duration_secs': 0.641766} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.005301] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1301.005621] env[62552]: INFO nova.compute.manager [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Took 7.09 seconds to spawn the instance on the hypervisor. [ 1301.005908] env[62552]: DEBUG nova.compute.manager [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1301.007040] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ceb3c8-a34f-4515-8690-0172223e2ed0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.526418] env[62552]: INFO nova.compute.manager [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Took 11.76 seconds to build instance. [ 1302.028705] env[62552]: DEBUG oslo_concurrency.lockutils [None req-9763c08d-d1cc-486c-8366-d32c5994bdd8 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "a45dba31-8b95-458e-a40d-dae81b26c27a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.267s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1302.385005] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "30e7c629-71b7-4430-8cc5-3460ecde577e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1302.385262] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "30e7c629-71b7-4430-8cc5-3460ecde577e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1302.887960] env[62552]: DEBUG nova.compute.manager [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1303.414328] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1303.414594] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1303.416099] env[62552]: INFO nova.compute.claims [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1304.468889] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39886e60-58d0-4db8-8509-5548afd1b0d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.476421] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e54143-fb96-4610-9dd9-9cb2c1b08093 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.504697] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e93afa-8a9e-4a32-aa13-371a206280a4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.511150] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90082224-4b10-4cdc-a942-83992b74a282 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.523523] env[62552]: DEBUG nova.compute.provider_tree [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1305.026244] env[62552]: DEBUG nova.scheduler.client.report [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1305.532806] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.118s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1305.533376] env[62552]: DEBUG nova.compute.manager [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1306.038139] env[62552]: DEBUG nova.compute.utils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1306.039529] env[62552]: DEBUG nova.compute.manager [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1306.039700] env[62552]: DEBUG nova.network.neutron [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1306.087135] env[62552]: DEBUG nova.policy [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7bc1a611cb24ab0bf7e2e046646039c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e4d140b840b453abcd828d2129de615', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1306.333141] env[62552]: DEBUG nova.network.neutron [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Successfully created port: 8104d478-ca45-4a14-ad2a-1c64f542177b {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1306.543618] env[62552]: DEBUG nova.compute.manager [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1307.554290] env[62552]: DEBUG nova.compute.manager [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1307.580713] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1307.581024] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1307.581198] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1307.581386] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1307.581538] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1307.581717] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1307.581971] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1307.582081] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1307.582254] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1307.582421] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1307.582596] env[62552]: DEBUG nova.virt.hardware [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1307.583491] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6e4068-ef02-4e2c-a097-99f549e239a2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.591285] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec1cf58-fbaf-4b80-9f0e-e07b79293e3a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.696886] env[62552]: DEBUG nova.compute.manager [req-5efaf782-8ad6-45c0-9da5-48b545e33392 req-fb108b58-fb82-46c5-aa8d-8a02e2e193c1 service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Received event network-vif-plugged-8104d478-ca45-4a14-ad2a-1c64f542177b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1307.697131] env[62552]: DEBUG oslo_concurrency.lockutils [req-5efaf782-8ad6-45c0-9da5-48b545e33392 req-fb108b58-fb82-46c5-aa8d-8a02e2e193c1 service nova] Acquiring lock "30e7c629-71b7-4430-8cc5-3460ecde577e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1307.697400] env[62552]: DEBUG oslo_concurrency.lockutils [req-5efaf782-8ad6-45c0-9da5-48b545e33392 req-fb108b58-fb82-46c5-aa8d-8a02e2e193c1 service nova] Lock "30e7c629-71b7-4430-8cc5-3460ecde577e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1307.697608] env[62552]: DEBUG oslo_concurrency.lockutils [req-5efaf782-8ad6-45c0-9da5-48b545e33392 req-fb108b58-fb82-46c5-aa8d-8a02e2e193c1 service nova] Lock "30e7c629-71b7-4430-8cc5-3460ecde577e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1307.697749] env[62552]: DEBUG nova.compute.manager [req-5efaf782-8ad6-45c0-9da5-48b545e33392 req-fb108b58-fb82-46c5-aa8d-8a02e2e193c1 service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] No waiting events found dispatching network-vif-plugged-8104d478-ca45-4a14-ad2a-1c64f542177b {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1307.697923] env[62552]: WARNING nova.compute.manager [req-5efaf782-8ad6-45c0-9da5-48b545e33392 req-fb108b58-fb82-46c5-aa8d-8a02e2e193c1 service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Received unexpected event network-vif-plugged-8104d478-ca45-4a14-ad2a-1c64f542177b for instance with vm_state building and task_state spawning. [ 1307.775518] env[62552]: DEBUG nova.network.neutron [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Successfully updated port: 8104d478-ca45-4a14-ad2a-1c64f542177b {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1308.278566] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "refresh_cache-30e7c629-71b7-4430-8cc5-3460ecde577e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1308.278720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "refresh_cache-30e7c629-71b7-4430-8cc5-3460ecde577e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1308.278870] env[62552]: DEBUG nova.network.neutron [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1308.810081] env[62552]: DEBUG nova.network.neutron [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1308.931703] env[62552]: DEBUG nova.network.neutron [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Updating instance_info_cache with network_info: [{"id": "8104d478-ca45-4a14-ad2a-1c64f542177b", "address": "fa:16:3e:a5:9c:e1", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8104d478-ca", "ovs_interfaceid": "8104d478-ca45-4a14-ad2a-1c64f542177b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1309.434904] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "refresh_cache-30e7c629-71b7-4430-8cc5-3460ecde577e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1309.435256] env[62552]: DEBUG nova.compute.manager [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Instance network_info: |[{"id": "8104d478-ca45-4a14-ad2a-1c64f542177b", "address": "fa:16:3e:a5:9c:e1", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8104d478-ca", "ovs_interfaceid": "8104d478-ca45-4a14-ad2a-1c64f542177b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1309.435751] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:9c:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8104d478-ca45-4a14-ad2a-1c64f542177b', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1309.443150] env[62552]: DEBUG oslo.service.loopingcall [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1309.443376] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1309.443974] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5fedcc15-0596-463d-8ec1-ecdb443c9bb3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.463830] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1309.463830] env[62552]: value = "task-1240280" [ 1309.463830] env[62552]: _type = "Task" [ 1309.463830] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.472780] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240280, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.722898] env[62552]: DEBUG nova.compute.manager [req-39d13053-7820-473e-aae4-b687310ae950 req-76f019c3-73f7-4f25-821d-4eaa11bebc34 service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Received event network-changed-8104d478-ca45-4a14-ad2a-1c64f542177b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1309.723171] env[62552]: DEBUG nova.compute.manager [req-39d13053-7820-473e-aae4-b687310ae950 req-76f019c3-73f7-4f25-821d-4eaa11bebc34 service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Refreshing instance network info cache due to event network-changed-8104d478-ca45-4a14-ad2a-1c64f542177b. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1309.723298] env[62552]: DEBUG oslo_concurrency.lockutils [req-39d13053-7820-473e-aae4-b687310ae950 req-76f019c3-73f7-4f25-821d-4eaa11bebc34 service nova] Acquiring lock "refresh_cache-30e7c629-71b7-4430-8cc5-3460ecde577e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1309.723449] env[62552]: DEBUG oslo_concurrency.lockutils [req-39d13053-7820-473e-aae4-b687310ae950 req-76f019c3-73f7-4f25-821d-4eaa11bebc34 service nova] Acquired lock "refresh_cache-30e7c629-71b7-4430-8cc5-3460ecde577e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1309.723616] env[62552]: DEBUG nova.network.neutron [req-39d13053-7820-473e-aae4-b687310ae950 req-76f019c3-73f7-4f25-821d-4eaa11bebc34 service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Refreshing network info cache for port 8104d478-ca45-4a14-ad2a-1c64f542177b {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1309.974061] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240280, 'name': CreateVM_Task, 'duration_secs': 0.365672} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.974387] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1309.974889] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1309.975076] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1309.975400] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1309.975671] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ec02a81-e743-427c-ae90-f74b61f358dd {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.979981] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1309.979981] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e0dc96-3619-072d-46be-ea6574a0bfb5" [ 1309.979981] env[62552]: _type = "Task" [ 1309.979981] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.987086] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e0dc96-3619-072d-46be-ea6574a0bfb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.404207] env[62552]: DEBUG nova.network.neutron [req-39d13053-7820-473e-aae4-b687310ae950 req-76f019c3-73f7-4f25-821d-4eaa11bebc34 service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Updated VIF entry in instance network info cache for port 8104d478-ca45-4a14-ad2a-1c64f542177b. {{(pid=62552) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1310.404571] env[62552]: DEBUG nova.network.neutron [req-39d13053-7820-473e-aae4-b687310ae950 req-76f019c3-73f7-4f25-821d-4eaa11bebc34 service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Updating instance_info_cache with network_info: [{"id": "8104d478-ca45-4a14-ad2a-1c64f542177b", "address": "fa:16:3e:a5:9c:e1", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8104d478-ca", "ovs_interfaceid": "8104d478-ca45-4a14-ad2a-1c64f542177b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1310.489553] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e0dc96-3619-072d-46be-ea6574a0bfb5, 'name': SearchDatastore_Task, 'duration_secs': 0.009529} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.489862] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1310.490121] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1310.490360] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1310.490509] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1310.490690] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1310.490942] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e31191a-9337-448d-b933-8833da516b52 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.498486] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1310.498654] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1310.499377] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-183e937d-7a7e-481d-8a9e-0c08bee0ae1d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.504078] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1310.504078] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5268d375-163b-ce93-1d57-92df16237159" [ 1310.504078] env[62552]: _type = "Task" [ 1310.504078] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.510874] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5268d375-163b-ce93-1d57-92df16237159, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.906991] env[62552]: DEBUG oslo_concurrency.lockutils [req-39d13053-7820-473e-aae4-b687310ae950 req-76f019c3-73f7-4f25-821d-4eaa11bebc34 service nova] Releasing lock "refresh_cache-30e7c629-71b7-4430-8cc5-3460ecde577e" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.014340] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]5268d375-163b-ce93-1d57-92df16237159, 'name': SearchDatastore_Task, 'duration_secs': 0.0075} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.015100] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0aad34c-91af-4993-ae36-8b1c1ab90f7c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.020204] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1311.020204] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c88db4-195c-210a-3a7a-409c1960b565" [ 1311.020204] env[62552]: _type = "Task" [ 1311.020204] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.027111] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c88db4-195c-210a-3a7a-409c1960b565, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.530875] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52c88db4-195c-210a-3a7a-409c1960b565, 'name': SearchDatastore_Task, 'duration_secs': 0.008888} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.531158] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.531419] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 30e7c629-71b7-4430-8cc5-3460ecde577e/30e7c629-71b7-4430-8cc5-3460ecde577e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1311.532044] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a317a02-5dc8-4fb3-a7e6-2f0e00a3b356 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.538266] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1311.538266] env[62552]: value = "task-1240281" [ 1311.538266] env[62552]: _type = "Task" [ 1311.538266] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.545372] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240281, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.048379] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240281, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474384} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.048801] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 30e7c629-71b7-4430-8cc5-3460ecde577e/30e7c629-71b7-4430-8cc5-3460ecde577e.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1312.048876] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1312.049099] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f913361-689c-4aba-8b9a-1d22ac42e98f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.054873] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1312.054873] env[62552]: value = "task-1240282" [ 1312.054873] env[62552]: _type = "Task" [ 1312.054873] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.061798] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240282, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.564505] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240282, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056981} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.564795] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1312.565570] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0565d5-192d-451a-aa5c-f1b4721719d7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.587654] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] 30e7c629-71b7-4430-8cc5-3460ecde577e/30e7c629-71b7-4430-8cc5-3460ecde577e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1312.587871] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8da070a2-b1e0-4634-9015-c0444d8dcd57 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.606338] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1312.606338] env[62552]: value = "task-1240283" [ 1312.606338] env[62552]: _type = "Task" [ 1312.606338] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.613638] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240283, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.116239] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240283, 'name': ReconfigVM_Task, 'duration_secs': 0.243655} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.116598] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Reconfigured VM instance instance-00000073 to attach disk [datastore2] 30e7c629-71b7-4430-8cc5-3460ecde577e/30e7c629-71b7-4430-8cc5-3460ecde577e.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1313.117170] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5cdc3b86-d8a4-4a0a-9e73-ddfeb0ca3e53 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.123204] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1313.123204] env[62552]: value = "task-1240284" [ 1313.123204] env[62552]: _type = "Task" [ 1313.123204] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.130172] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240284, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.632659] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240284, 'name': Rename_Task, 'duration_secs': 0.136204} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.632926] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1313.633183] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14198dfb-db1b-4ab5-acad-5340834b7adf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.638851] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1313.638851] env[62552]: value = "task-1240285" [ 1313.638851] env[62552]: _type = "Task" [ 1313.638851] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.645892] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240285, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.148695] env[62552]: DEBUG oslo_vmware.api [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240285, 'name': PowerOnVM_Task, 'duration_secs': 0.426461} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.149073] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1314.149195] env[62552]: INFO nova.compute.manager [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Took 6.59 seconds to spawn the instance on the hypervisor. [ 1314.149302] env[62552]: DEBUG nova.compute.manager [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1314.150034] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb76c7e-35b2-4370-b75c-ec934b4e4d14 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.666993] env[62552]: INFO nova.compute.manager [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Took 11.27 seconds to build instance. [ 1315.169737] env[62552]: DEBUG oslo_concurrency.lockutils [None req-eaa5dee8-4893-4e81-9888-da28fea41920 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "30e7c629-71b7-4430-8cc5-3460ecde577e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.784s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1315.675088] env[62552]: DEBUG oslo_concurrency.lockutils [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "30e7c629-71b7-4430-8cc5-3460ecde577e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1315.675407] env[62552]: DEBUG oslo_concurrency.lockutils [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "30e7c629-71b7-4430-8cc5-3460ecde577e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1315.675632] env[62552]: DEBUG oslo_concurrency.lockutils [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "30e7c629-71b7-4430-8cc5-3460ecde577e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1315.675877] env[62552]: DEBUG oslo_concurrency.lockutils [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "30e7c629-71b7-4430-8cc5-3460ecde577e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1315.676078] env[62552]: DEBUG oslo_concurrency.lockutils [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "30e7c629-71b7-4430-8cc5-3460ecde577e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1315.678789] env[62552]: INFO nova.compute.manager [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Terminating instance [ 1316.182625] env[62552]: DEBUG nova.compute.manager [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1316.183041] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1316.184223] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb9e436-d3a9-49d8-9fcb-29640a2d06bf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.191573] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1316.191807] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba57f45a-e8c6-4bae-8943-b5e077d8a5b1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.196987] env[62552]: DEBUG oslo_vmware.api [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1316.196987] env[62552]: value = "task-1240286" [ 1316.196987] env[62552]: _type = "Task" [ 1316.196987] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.205619] env[62552]: DEBUG oslo_vmware.api [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.708648] env[62552]: DEBUG oslo_vmware.api [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240286, 'name': PowerOffVM_Task, 'duration_secs': 0.223672} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1316.708921] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1316.709111] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1316.709369] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aeaa530a-688f-4e2f-8e89-e84b08314603 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.774667] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1316.774667] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1316.774667] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleting the datastore file [datastore2] 30e7c629-71b7-4430-8cc5-3460ecde577e {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1316.774921] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d83307fe-e1ed-4e9f-8deb-45a43f65911c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.781230] env[62552]: DEBUG oslo_vmware.api [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1316.781230] env[62552]: value = "task-1240288" [ 1316.781230] env[62552]: _type = "Task" [ 1316.781230] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.788794] env[62552]: DEBUG oslo_vmware.api [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240288, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.290870] env[62552]: DEBUG oslo_vmware.api [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240288, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13181} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.293306] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1317.293306] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1317.293306] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1317.293306] env[62552]: INFO nova.compute.manager [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1317.293306] env[62552]: DEBUG oslo.service.loopingcall [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1317.293306] env[62552]: DEBUG nova.compute.manager [-] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1317.293306] env[62552]: DEBUG nova.network.neutron [-] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1317.618098] env[62552]: DEBUG nova.compute.manager [req-34d540b9-d7b6-484c-b0cc-f8685fd4a14f req-56072f2c-0194-477b-93e3-a89cefa9f4eb service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Received event network-vif-deleted-8104d478-ca45-4a14-ad2a-1c64f542177b {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1317.618347] env[62552]: INFO nova.compute.manager [req-34d540b9-d7b6-484c-b0cc-f8685fd4a14f req-56072f2c-0194-477b-93e3-a89cefa9f4eb service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Neutron deleted interface 8104d478-ca45-4a14-ad2a-1c64f542177b; detaching it from the instance and deleting it from the info cache [ 1317.618470] env[62552]: DEBUG nova.network.neutron [req-34d540b9-d7b6-484c-b0cc-f8685fd4a14f req-56072f2c-0194-477b-93e3-a89cefa9f4eb service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1318.095477] env[62552]: DEBUG nova.network.neutron [-] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1318.120632] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c0e8587-ebe3-4156-aa25-a8de7dc64791 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.130533] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e4e856-1392-453e-ab70-de6e7aaa2d96 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.153799] env[62552]: DEBUG nova.compute.manager [req-34d540b9-d7b6-484c-b0cc-f8685fd4a14f req-56072f2c-0194-477b-93e3-a89cefa9f4eb service nova] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Detach interface failed, port_id=8104d478-ca45-4a14-ad2a-1c64f542177b, reason: Instance 30e7c629-71b7-4430-8cc5-3460ecde577e could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1318.598259] env[62552]: INFO nova.compute.manager [-] [instance: 30e7c629-71b7-4430-8cc5-3460ecde577e] Took 1.31 seconds to deallocate network for instance. [ 1319.104497] env[62552]: DEBUG oslo_concurrency.lockutils [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1319.104835] env[62552]: DEBUG oslo_concurrency.lockutils [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1319.105117] env[62552]: DEBUG nova.objects.instance [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lazy-loading 'resources' on Instance uuid 30e7c629-71b7-4430-8cc5-3460ecde577e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1319.728840] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea47caf2-8098-4634-ad54-e7476f18f7d9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.736229] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e06c8e-9316-473b-b3b1-90e9e1af8fc6 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.766410] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241a1c3f-cb1d-441f-a877-485f090dc7bc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.773264] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80d75e2-a6d5-41b5-9b8d-3fa98416215e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.785867] env[62552]: DEBUG nova.compute.provider_tree [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1320.289382] env[62552]: DEBUG nova.scheduler.client.report [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1320.794549] env[62552]: DEBUG oslo_concurrency.lockutils [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.690s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1320.814703] env[62552]: INFO nova.scheduler.client.report [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted allocations for instance 30e7c629-71b7-4430-8cc5-3460ecde577e [ 1321.322558] env[62552]: DEBUG oslo_concurrency.lockutils [None req-102f51e6-f2ec-4314-9376-d7f521cc674a tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "30e7c629-71b7-4430-8cc5-3460ecde577e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.647s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1322.288126] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "a45dba31-8b95-458e-a40d-dae81b26c27a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1322.288519] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "a45dba31-8b95-458e-a40d-dae81b26c27a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1322.288826] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "a45dba31-8b95-458e-a40d-dae81b26c27a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1322.289120] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "a45dba31-8b95-458e-a40d-dae81b26c27a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1322.289367] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "a45dba31-8b95-458e-a40d-dae81b26c27a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1322.291638] env[62552]: INFO nova.compute.manager [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Terminating instance [ 1322.795732] env[62552]: DEBUG nova.compute.manager [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1322.796026] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1322.796990] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfa1618-c870-450c-a283-9613da8a3986 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.804713] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1322.804939] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-916ff7b9-7124-4b2a-9ebf-c99133a3dccb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.810919] env[62552]: DEBUG oslo_vmware.api [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1322.810919] env[62552]: value = "task-1240289" [ 1322.810919] env[62552]: _type = "Task" [ 1322.810919] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.819383] env[62552]: DEBUG oslo_vmware.api [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240289, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.320342] env[62552]: DEBUG oslo_vmware.api [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240289, 'name': PowerOffVM_Task, 'duration_secs': 0.18542} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.320729] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1323.320773] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1323.320989] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce105033-d919-4a25-b6b2-56b476678f1a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.380905] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1323.381145] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1323.381339] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleting the datastore file [datastore2] a45dba31-8b95-458e-a40d-dae81b26c27a {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1323.381598] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb039ddf-8328-4d3e-95f4-0a1f850db809 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.387720] env[62552]: DEBUG oslo_vmware.api [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1323.387720] env[62552]: value = "task-1240291" [ 1323.387720] env[62552]: _type = "Task" [ 1323.387720] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.394588] env[62552]: DEBUG oslo_vmware.api [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.897569] env[62552]: DEBUG oslo_vmware.api [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124034} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.897828] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1323.898021] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1323.898218] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1323.898398] env[62552]: INFO nova.compute.manager [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1323.898637] env[62552]: DEBUG oslo.service.loopingcall [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1323.898830] env[62552]: DEBUG nova.compute.manager [-] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1323.898929] env[62552]: DEBUG nova.network.neutron [-] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1324.128429] env[62552]: DEBUG nova.compute.manager [req-50072ce3-0e85-4be7-8770-f624d39e9ef7 req-0111a48f-323e-4dd2-a954-1dbfc102c530 service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Received event network-vif-deleted-5a32c6a9-da6a-46be-9196-ac1d7554ad20 {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1324.128539] env[62552]: INFO nova.compute.manager [req-50072ce3-0e85-4be7-8770-f624d39e9ef7 req-0111a48f-323e-4dd2-a954-1dbfc102c530 service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Neutron deleted interface 5a32c6a9-da6a-46be-9196-ac1d7554ad20; detaching it from the instance and deleting it from the info cache [ 1324.128705] env[62552]: DEBUG nova.network.neutron [req-50072ce3-0e85-4be7-8770-f624d39e9ef7 req-0111a48f-323e-4dd2-a954-1dbfc102c530 service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1324.608350] env[62552]: DEBUG nova.network.neutron [-] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1324.631501] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b406445-1584-440a-9c53-895fa9d20d15 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.641079] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea499d27-2d3a-4c47-bfe6-d5e2f8a6a7a9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.662717] env[62552]: DEBUG nova.compute.manager [req-50072ce3-0e85-4be7-8770-f624d39e9ef7 req-0111a48f-323e-4dd2-a954-1dbfc102c530 service nova] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Detach interface failed, port_id=5a32c6a9-da6a-46be-9196-ac1d7554ad20, reason: Instance a45dba31-8b95-458e-a40d-dae81b26c27a could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1325.111025] env[62552]: INFO nova.compute.manager [-] [instance: a45dba31-8b95-458e-a40d-dae81b26c27a] Took 1.21 seconds to deallocate network for instance. [ 1325.617745] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.618105] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.618314] env[62552]: DEBUG nova.objects.instance [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lazy-loading 'resources' on Instance uuid a45dba31-8b95-458e-a40d-dae81b26c27a {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1326.161594] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96f8270-bad5-4bd6-a49a-2b9b02eb7647 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.169388] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32bc397-94e6-4993-a14d-97fdd6f5b3ce {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.199647] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c888e05-d0b4-4d09-8fe9-fbb54e067f87 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.206566] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff6b446-9816-46b1-8875-d82d9cdb0043 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.220894] env[62552]: DEBUG nova.compute.provider_tree [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1326.723845] env[62552]: DEBUG nova.scheduler.client.report [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1327.228935] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1327.249198] env[62552]: INFO nova.scheduler.client.report [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted allocations for instance a45dba31-8b95-458e-a40d-dae81b26c27a [ 1327.759377] env[62552]: DEBUG oslo_concurrency.lockutils [None req-c170e4c3-cb3c-4bf5-83b4-73177f280b78 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "a45dba31-8b95-458e-a40d-dae81b26c27a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.471s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1329.196476] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1329.196781] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1329.699054] env[62552]: DEBUG nova.compute.manager [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1330.219526] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.219785] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.221316] env[62552]: INFO nova.compute.claims [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1331.266152] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ae5bc2-f11a-452c-9475-f2a7eee5e81d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.274217] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2460d3d-19e1-4893-8839-cd3de4288533 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.302592] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfc6cf5-c429-4fa1-9364-5eebeecb014a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.309321] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680071e7-c22a-416c-aa5b-1482ff9fce81 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.321832] env[62552]: DEBUG nova.compute.provider_tree [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1331.825036] env[62552]: DEBUG nova.scheduler.client.report [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1332.329917] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.110s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1332.330312] env[62552]: DEBUG nova.compute.manager [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1332.836179] env[62552]: DEBUG nova.compute.utils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1332.837932] env[62552]: DEBUG nova.compute.manager [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1332.838119] env[62552]: DEBUG nova.network.neutron [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1332.895272] env[62552]: DEBUG nova.policy [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7bc1a611cb24ab0bf7e2e046646039c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e4d140b840b453abcd828d2129de615', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1333.139436] env[62552]: DEBUG nova.network.neutron [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Successfully created port: 05b1e4e4-3cce-4f97-bf49-a81e6743d08a {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1333.341625] env[62552]: DEBUG nova.compute.manager [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1334.351102] env[62552]: DEBUG nova.compute.manager [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1334.377370] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1334.377589] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1334.377749] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1334.377932] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1334.378101] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1334.378257] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1334.378463] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1334.378625] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1334.378790] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1334.378977] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1334.379137] env[62552]: DEBUG nova.virt.hardware [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1334.379997] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899795e0-3b36-4092-9d96-9c1d37c4223f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.387648] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91005e65-a318-4e81-ac90-c28029cc8aeb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.504183] env[62552]: DEBUG nova.compute.manager [req-e8eeafa4-94c2-47d3-9dcc-9ebb4e8c998e req-f57e88b7-0a5b-4725-bbac-ce19c3b6e74c service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Received event network-vif-plugged-05b1e4e4-3cce-4f97-bf49-a81e6743d08a {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1334.504413] env[62552]: DEBUG oslo_concurrency.lockutils [req-e8eeafa4-94c2-47d3-9dcc-9ebb4e8c998e req-f57e88b7-0a5b-4725-bbac-ce19c3b6e74c service nova] Acquiring lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1334.504628] env[62552]: DEBUG oslo_concurrency.lockutils [req-e8eeafa4-94c2-47d3-9dcc-9ebb4e8c998e req-f57e88b7-0a5b-4725-bbac-ce19c3b6e74c service nova] Lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1334.504856] env[62552]: DEBUG oslo_concurrency.lockutils [req-e8eeafa4-94c2-47d3-9dcc-9ebb4e8c998e req-f57e88b7-0a5b-4725-bbac-ce19c3b6e74c service nova] Lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1334.504970] env[62552]: DEBUG nova.compute.manager [req-e8eeafa4-94c2-47d3-9dcc-9ebb4e8c998e req-f57e88b7-0a5b-4725-bbac-ce19c3b6e74c service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] No waiting events found dispatching network-vif-plugged-05b1e4e4-3cce-4f97-bf49-a81e6743d08a {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1334.505146] env[62552]: WARNING nova.compute.manager [req-e8eeafa4-94c2-47d3-9dcc-9ebb4e8c998e req-f57e88b7-0a5b-4725-bbac-ce19c3b6e74c service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Received unexpected event network-vif-plugged-05b1e4e4-3cce-4f97-bf49-a81e6743d08a for instance with vm_state building and task_state spawning. [ 1335.047500] env[62552]: DEBUG nova.network.neutron [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Successfully updated port: 05b1e4e4-3cce-4f97-bf49-a81e6743d08a {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1335.070594] env[62552]: DEBUG nova.compute.manager [req-f62c60b6-9872-45cd-bf55-2d362cda735d req-381133ff-8454-4b4b-8d41-30f39152b651 service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Received event network-changed-05b1e4e4-3cce-4f97-bf49-a81e6743d08a {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1335.070806] env[62552]: DEBUG nova.compute.manager [req-f62c60b6-9872-45cd-bf55-2d362cda735d req-381133ff-8454-4b4b-8d41-30f39152b651 service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Refreshing instance network info cache due to event network-changed-05b1e4e4-3cce-4f97-bf49-a81e6743d08a. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1335.071037] env[62552]: DEBUG oslo_concurrency.lockutils [req-f62c60b6-9872-45cd-bf55-2d362cda735d req-381133ff-8454-4b4b-8d41-30f39152b651 service nova] Acquiring lock "refresh_cache-126c6459-e37f-4a2a-9a8c-3c2a911d9653" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1335.071191] env[62552]: DEBUG oslo_concurrency.lockutils [req-f62c60b6-9872-45cd-bf55-2d362cda735d req-381133ff-8454-4b4b-8d41-30f39152b651 service nova] Acquired lock "refresh_cache-126c6459-e37f-4a2a-9a8c-3c2a911d9653" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1335.071358] env[62552]: DEBUG nova.network.neutron [req-f62c60b6-9872-45cd-bf55-2d362cda735d req-381133ff-8454-4b4b-8d41-30f39152b651 service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Refreshing network info cache for port 05b1e4e4-3cce-4f97-bf49-a81e6743d08a {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1335.549808] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "refresh_cache-126c6459-e37f-4a2a-9a8c-3c2a911d9653" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1335.601196] env[62552]: DEBUG nova.network.neutron [req-f62c60b6-9872-45cd-bf55-2d362cda735d req-381133ff-8454-4b4b-8d41-30f39152b651 service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1335.673934] env[62552]: DEBUG nova.network.neutron [req-f62c60b6-9872-45cd-bf55-2d362cda735d req-381133ff-8454-4b4b-8d41-30f39152b651 service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1336.176706] env[62552]: DEBUG oslo_concurrency.lockutils [req-f62c60b6-9872-45cd-bf55-2d362cda735d req-381133ff-8454-4b4b-8d41-30f39152b651 service nova] Releasing lock "refresh_cache-126c6459-e37f-4a2a-9a8c-3c2a911d9653" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1336.177125] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "refresh_cache-126c6459-e37f-4a2a-9a8c-3c2a911d9653" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1336.177294] env[62552]: DEBUG nova.network.neutron [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1336.707294] env[62552]: DEBUG nova.network.neutron [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1336.826664] env[62552]: DEBUG nova.network.neutron [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Updating instance_info_cache with network_info: [{"id": "05b1e4e4-3cce-4f97-bf49-a81e6743d08a", "address": "fa:16:3e:f1:c2:ee", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05b1e4e4-3c", "ovs_interfaceid": "05b1e4e4-3cce-4f97-bf49-a81e6743d08a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1337.329566] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "refresh_cache-126c6459-e37f-4a2a-9a8c-3c2a911d9653" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1337.329880] env[62552]: DEBUG nova.compute.manager [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Instance network_info: |[{"id": "05b1e4e4-3cce-4f97-bf49-a81e6743d08a", "address": "fa:16:3e:f1:c2:ee", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05b1e4e4-3c", "ovs_interfaceid": "05b1e4e4-3cce-4f97-bf49-a81e6743d08a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1337.330346] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:c2:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '05b1e4e4-3cce-4f97-bf49-a81e6743d08a', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1337.337736] env[62552]: DEBUG oslo.service.loopingcall [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1337.337943] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1337.338185] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c13b12b-7b46-42bd-aa24-b49b210ac975 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.358337] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1337.358337] env[62552]: value = "task-1240292" [ 1337.358337] env[62552]: _type = "Task" [ 1337.358337] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.365486] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240292, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.868469] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240292, 'name': CreateVM_Task, 'duration_secs': 0.299548} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.868823] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1337.869321] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1337.869490] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1337.869811] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1337.870071] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-812e6ccb-847b-4203-b73d-33821a94e8e9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.874217] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1337.874217] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520930c2-b454-135c-18fc-6d3042a48ae1" [ 1337.874217] env[62552]: _type = "Task" [ 1337.874217] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.881266] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520930c2-b454-135c-18fc-6d3042a48ae1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.384212] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520930c2-b454-135c-18fc-6d3042a48ae1, 'name': SearchDatastore_Task, 'duration_secs': 0.009645} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.384519] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1338.384775] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1338.385018] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1338.385179] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1338.385365] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1338.385619] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65eeb6fe-e77e-45d5-9bee-0b7b846492b8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.393200] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1338.393370] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1338.394063] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1731c74-3159-4866-9af3-a66548832f88 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.398542] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1338.398542] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529c90d9-2c09-c4bf-171a-66f2deabd9fe" [ 1338.398542] env[62552]: _type = "Task" [ 1338.398542] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.405482] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529c90d9-2c09-c4bf-171a-66f2deabd9fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.909084] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]529c90d9-2c09-c4bf-171a-66f2deabd9fe, 'name': SearchDatastore_Task, 'duration_secs': 0.007831} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.909877] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bab1d8e0-be4d-4af8-aa31-5af9ec18b09b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.914640] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1338.914640] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e965e3-a0f9-a264-e449-e8e54ff7fc7b" [ 1338.914640] env[62552]: _type = "Task" [ 1338.914640] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.921517] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e965e3-a0f9-a264-e449-e8e54ff7fc7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.425040] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52e965e3-a0f9-a264-e449-e8e54ff7fc7b, 'name': SearchDatastore_Task, 'duration_secs': 0.008281} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.425313] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1339.425574] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 126c6459-e37f-4a2a-9a8c-3c2a911d9653/126c6459-e37f-4a2a-9a8c-3c2a911d9653.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1339.425825] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ee34ffe-bbbe-446c-afc4-98266756e5b9 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.432638] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1339.432638] env[62552]: value = "task-1240293" [ 1339.432638] env[62552]: _type = "Task" [ 1339.432638] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.443421] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.698433] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1339.942556] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240293, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.407509} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.942925] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] 126c6459-e37f-4a2a-9a8c-3c2a911d9653/126c6459-e37f-4a2a-9a8c-3c2a911d9653.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1339.943040] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1339.943295] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd0182f9-09ba-4493-9405-7eeda96152e1 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.949273] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1339.949273] env[62552]: value = "task-1240294" [ 1339.949273] env[62552]: _type = "Task" [ 1339.949273] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.955734] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240294, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.458966] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240294, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06505} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.459269] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1340.460454] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b69c65-0b81-4963-b30a-4f151e804835 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.482715] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] 126c6459-e37f-4a2a-9a8c-3c2a911d9653/126c6459-e37f-4a2a-9a8c-3c2a911d9653.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1340.482968] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f34cabb7-7bf6-48a3-be97-c2f948fbf251 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.501389] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1340.501389] env[62552]: value = "task-1240295" [ 1340.501389] env[62552]: _type = "Task" [ 1340.501389] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.508635] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240295, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.698286] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1340.698483] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Starting heal instance info cache {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1341.010809] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240295, 'name': ReconfigVM_Task, 'duration_secs': 0.249338} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.011213] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Reconfigured VM instance instance-00000074 to attach disk [datastore2] 126c6459-e37f-4a2a-9a8c-3c2a911d9653/126c6459-e37f-4a2a-9a8c-3c2a911d9653.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1341.011805] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6b1bf95-ec2d-4ede-857e-39b117e84ce2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.017399] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1341.017399] env[62552]: value = "task-1240296" [ 1341.017399] env[62552]: _type = "Task" [ 1341.017399] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.024131] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240296, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.201280] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Didn't find any instances for network info cache update. {{(pid=62552) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10357}} [ 1341.201534] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.527732] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240296, 'name': Rename_Task, 'duration_secs': 0.144244} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.528059] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1341.528321] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f80a175f-265b-43fc-9a2e-46cd878ce921 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.534157] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1341.534157] env[62552]: value = "task-1240297" [ 1341.534157] env[62552]: _type = "Task" [ 1341.534157] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.542086] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240297, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.697949] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1342.044443] env[62552]: DEBUG oslo_vmware.api [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240297, 'name': PowerOnVM_Task, 'duration_secs': 0.454089} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.044812] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1342.044914] env[62552]: INFO nova.compute.manager [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Took 7.69 seconds to spawn the instance on the hypervisor. [ 1342.045121] env[62552]: DEBUG nova.compute.manager [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1342.045866] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b610829-ec29-451c-8229-e745c52ab3aa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.561183] env[62552]: INFO nova.compute.manager [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Took 12.36 seconds to build instance. [ 1342.693361] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1342.697973] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1343.063455] env[62552]: DEBUG oslo_concurrency.lockutils [None req-e6c8c705-4779-4137-a302-d67f2ba9468c tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.867s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1343.660506] env[62552]: DEBUG oslo_concurrency.lockutils [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1343.660739] env[62552]: DEBUG oslo_concurrency.lockutils [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1343.660952] env[62552]: DEBUG oslo_concurrency.lockutils [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1343.661158] env[62552]: DEBUG oslo_concurrency.lockutils [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1343.661378] env[62552]: DEBUG oslo_concurrency.lockutils [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1343.663617] env[62552]: INFO nova.compute.manager [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Terminating instance [ 1343.697249] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.167588] env[62552]: DEBUG nova.compute.manager [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1344.167967] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1344.168759] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ff068b-29e0-480f-aeb7-d9c054caeb4f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.178171] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1344.178418] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0e667fa-02a3-42d0-8e2e-2a66736f1ddf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.184528] env[62552]: DEBUG oslo_vmware.api [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1344.184528] env[62552]: value = "task-1240298" [ 1344.184528] env[62552]: _type = "Task" [ 1344.184528] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.192094] env[62552]: DEBUG oslo_vmware.api [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240298, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.200042] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1344.200248] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1344.200410] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1344.200561] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62552) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1344.201305] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c298a85-e8d1-4557-a2d1-d4794964cbdb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.208841] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcba8137-df4b-4123-a784-3e7ee3bf4dcf {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.224042] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55195b81-318a-4de9-b934-d97ba93708e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.231223] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb401b77-ea33-4f56-a3c9-5818b72f4a68 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.261694] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181301MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62552) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1344.261871] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1344.262101] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1344.694658] env[62552]: DEBUG oslo_vmware.api [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240298, 'name': PowerOffVM_Task, 'duration_secs': 0.178263} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.695010] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1344.695225] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1344.695473] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b03a4209-e92f-4b20-bcce-1068e83df99b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.768476] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1344.768651] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1344.768822] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleting the datastore file [datastore2] 126c6459-e37f-4a2a-9a8c-3c2a911d9653 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1344.769295] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e319d44-f666-4c80-893b-76e343308a21 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.775135] env[62552]: DEBUG oslo_vmware.api [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1344.775135] env[62552]: value = "task-1240300" [ 1344.775135] env[62552]: _type = "Task" [ 1344.775135] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.783757] env[62552]: DEBUG oslo_vmware.api [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240300, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.285665] env[62552]: DEBUG oslo_vmware.api [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240300, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149672} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.286089] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1345.286139] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1345.286330] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1345.286512] env[62552]: INFO nova.compute.manager [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1345.286758] env[62552]: DEBUG oslo.service.loopingcall [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1345.286953] env[62552]: DEBUG nova.compute.manager [-] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1345.287067] env[62552]: DEBUG nova.network.neutron [-] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1345.289863] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 604e8472-7af4-49a1-9a0c-459482db9d6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1345.290033] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Instance 126c6459-e37f-4a2a-9a8c-3c2a911d9653 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62552) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1345.290181] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1345.290317] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62552) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1345.329716] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0f4f1b-6606-4eed-bc05-f7d7c0202972 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.338079] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fc971f-d443-4a2d-87ef-a6835be68a05 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.377033] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f91d08d-e77e-43ae-b49a-539e682e0c40 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.382300] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98745e9d-bc88-4caf-9e7d-228e3c882c1d {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.402173] env[62552]: DEBUG nova.compute.provider_tree [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1345.628704] env[62552]: DEBUG nova.compute.manager [req-26a4df85-66e9-4107-85bb-9fbdbce1ec11 req-332575ba-e1cf-4a0f-b6ce-b9451572c516 service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Received event network-vif-deleted-05b1e4e4-3cce-4f97-bf49-a81e6743d08a {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1345.628924] env[62552]: INFO nova.compute.manager [req-26a4df85-66e9-4107-85bb-9fbdbce1ec11 req-332575ba-e1cf-4a0f-b6ce-b9451572c516 service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Neutron deleted interface 05b1e4e4-3cce-4f97-bf49-a81e6743d08a; detaching it from the instance and deleting it from the info cache [ 1345.629122] env[62552]: DEBUG nova.network.neutron [req-26a4df85-66e9-4107-85bb-9fbdbce1ec11 req-332575ba-e1cf-4a0f-b6ce-b9451572c516 service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1345.905997] env[62552]: DEBUG nova.scheduler.client.report [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1346.106553] env[62552]: DEBUG nova.network.neutron [-] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1346.131638] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1834e4cf-7cfe-4c6c-b536-f381e92bc649 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.141220] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaca20b-0585-4a39-90e5-ca211f8977d2 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.163116] env[62552]: DEBUG nova.compute.manager [req-26a4df85-66e9-4107-85bb-9fbdbce1ec11 req-332575ba-e1cf-4a0f-b6ce-b9451572c516 service nova] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Detach interface failed, port_id=05b1e4e4-3cce-4f97-bf49-a81e6743d08a, reason: Instance 126c6459-e37f-4a2a-9a8c-3c2a911d9653 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1346.410876] env[62552]: DEBUG nova.compute.resource_tracker [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62552) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1346.411246] env[62552]: DEBUG oslo_concurrency.lockutils [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.149s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.610011] env[62552]: INFO nova.compute.manager [-] [instance: 126c6459-e37f-4a2a-9a8c-3c2a911d9653] Took 1.32 seconds to deallocate network for instance. [ 1347.116298] env[62552]: DEBUG oslo_concurrency.lockutils [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1347.116572] env[62552]: DEBUG oslo_concurrency.lockutils [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1347.116862] env[62552]: DEBUG nova.objects.instance [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lazy-loading 'resources' on Instance uuid 126c6459-e37f-4a2a-9a8c-3c2a911d9653 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1347.412357] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1347.412694] env[62552]: DEBUG nova.compute.manager [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62552) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1347.660501] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0215120e-0af9-48d3-ae91-9d8cc0d10a3a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.667877] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1790a6f0-9a2a-458c-8046-adaa5be99d22 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.696348] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7a95b5-fbfe-4b01-b51f-1af34a770160 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.703260] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69066cd6-d94f-4727-992c-f5afe87d22df {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.716859] env[62552]: DEBUG nova.compute.provider_tree [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1348.220341] env[62552]: DEBUG nova.scheduler.client.report [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1348.698361] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1348.725479] env[62552]: DEBUG oslo_concurrency.lockutils [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1348.748188] env[62552]: INFO nova.scheduler.client.report [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted allocations for instance 126c6459-e37f-4a2a-9a8c-3c2a911d9653 [ 1349.256601] env[62552]: DEBUG oslo_concurrency.lockutils [None req-227ce14c-6258-483a-ae92-1363946e54ea tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "126c6459-e37f-4a2a-9a8c-3c2a911d9653" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.596s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1350.659060] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1350.659060] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1351.161718] env[62552]: DEBUG nova.compute.manager [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Starting instance... {{(pid=62552) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1351.683252] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1351.683559] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1351.685018] env[62552]: INFO nova.compute.claims [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1351.692948] env[62552]: DEBUG oslo_service.periodic_task [None req-13da2782-6d65-4e37-8efe-8534db2eb6c1 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62552) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.729173] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4139d5-e12d-4333-bab8-6e01e65e8b1c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.736660] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc002c4-29a2-47e2-89fb-b001e3f8b101 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.765927] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fea0812-0d80-4389-b52b-ce180d6291a5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.772476] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d6a845-d2b0-432b-978d-fcb3f0d969bb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.784886] env[62552]: DEBUG nova.compute.provider_tree [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1353.288046] env[62552]: DEBUG nova.scheduler.client.report [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1353.793553] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.110s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1353.794114] env[62552]: DEBUG nova.compute.manager [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Start building networks asynchronously for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1354.300046] env[62552]: DEBUG nova.compute.utils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Using /dev/sd instead of None {{(pid=62552) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1354.300860] env[62552]: DEBUG nova.compute.manager [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Allocating IP information in the background. {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1354.301044] env[62552]: DEBUG nova.network.neutron [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] allocate_for_instance() {{(pid=62552) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1354.354060] env[62552]: DEBUG nova.policy [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7bc1a611cb24ab0bf7e2e046646039c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e4d140b840b453abcd828d2129de615', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62552) authorize /opt/stack/nova/nova/policy.py:201}} [ 1354.594876] env[62552]: DEBUG nova.network.neutron [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Successfully created port: 68cc9439-cb9d-4f29-b1c7-048e9748adfd {{(pid=62552) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1354.803781] env[62552]: DEBUG nova.compute.manager [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Start building block device mappings for instance. {{(pid=62552) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1355.815609] env[62552]: DEBUG nova.compute.manager [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Start spawning the instance on the hypervisor. {{(pid=62552) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1355.841634] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T11:29:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T11:28:51Z,direct_url=,disk_format='vmdk',id=562b48de-e081-4c62-a3e2-cf4ccfd0d1ee,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='04a4549325ec47ddb8b99e2a28b02183',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T11:28:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1355.841948] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1355.842171] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image limits 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1355.842372] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Flavor pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1355.842582] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Image pref 0:0:0 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1355.842755] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62552) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1355.842969] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1355.843153] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1355.843320] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Got 1 possible topologies {{(pid=62552) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1355.843485] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1355.843680] env[62552]: DEBUG nova.virt.hardware [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62552) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1355.844741] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f612b1-f942-424e-bbf2-09a0d7981c6a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.852889] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f28715-7076-4be0-96f0-aaca7216a9f5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.936229] env[62552]: DEBUG nova.compute.manager [req-bcd13891-19fd-4fd2-a557-bdbd6a51044d req-91c7cc59-159e-455e-8273-4a228cbf69f1 service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Received event network-vif-plugged-68cc9439-cb9d-4f29-b1c7-048e9748adfd {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1355.936495] env[62552]: DEBUG oslo_concurrency.lockutils [req-bcd13891-19fd-4fd2-a557-bdbd6a51044d req-91c7cc59-159e-455e-8273-4a228cbf69f1 service nova] Acquiring lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1355.936685] env[62552]: DEBUG oslo_concurrency.lockutils [req-bcd13891-19fd-4fd2-a557-bdbd6a51044d req-91c7cc59-159e-455e-8273-4a228cbf69f1 service nova] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1355.936859] env[62552]: DEBUG oslo_concurrency.lockutils [req-bcd13891-19fd-4fd2-a557-bdbd6a51044d req-91c7cc59-159e-455e-8273-4a228cbf69f1 service nova] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1355.937042] env[62552]: DEBUG nova.compute.manager [req-bcd13891-19fd-4fd2-a557-bdbd6a51044d req-91c7cc59-159e-455e-8273-4a228cbf69f1 service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] No waiting events found dispatching network-vif-plugged-68cc9439-cb9d-4f29-b1c7-048e9748adfd {{(pid=62552) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1355.937246] env[62552]: WARNING nova.compute.manager [req-bcd13891-19fd-4fd2-a557-bdbd6a51044d req-91c7cc59-159e-455e-8273-4a228cbf69f1 service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Received unexpected event network-vif-plugged-68cc9439-cb9d-4f29-b1c7-048e9748adfd for instance with vm_state building and task_state spawning. [ 1356.560838] env[62552]: DEBUG nova.network.neutron [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Successfully updated port: 68cc9439-cb9d-4f29-b1c7-048e9748adfd {{(pid=62552) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1356.582797] env[62552]: DEBUG nova.compute.manager [req-1240d292-ad14-4531-b25d-020b90f7a1b5 req-57fc8a33-4c7e-43f8-a1d8-58c9b2be0109 service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Received event network-changed-68cc9439-cb9d-4f29-b1c7-048e9748adfd {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1356.583074] env[62552]: DEBUG nova.compute.manager [req-1240d292-ad14-4531-b25d-020b90f7a1b5 req-57fc8a33-4c7e-43f8-a1d8-58c9b2be0109 service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Refreshing instance network info cache due to event network-changed-68cc9439-cb9d-4f29-b1c7-048e9748adfd. {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1356.583269] env[62552]: DEBUG oslo_concurrency.lockutils [req-1240d292-ad14-4531-b25d-020b90f7a1b5 req-57fc8a33-4c7e-43f8-a1d8-58c9b2be0109 service nova] Acquiring lock "refresh_cache-d8763396-efbb-4e45-b9c8-efdcd8be27e5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1356.583423] env[62552]: DEBUG oslo_concurrency.lockutils [req-1240d292-ad14-4531-b25d-020b90f7a1b5 req-57fc8a33-4c7e-43f8-a1d8-58c9b2be0109 service nova] Acquired lock "refresh_cache-d8763396-efbb-4e45-b9c8-efdcd8be27e5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1356.583582] env[62552]: DEBUG nova.network.neutron [req-1240d292-ad14-4531-b25d-020b90f7a1b5 req-57fc8a33-4c7e-43f8-a1d8-58c9b2be0109 service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Refreshing network info cache for port 68cc9439-cb9d-4f29-b1c7-048e9748adfd {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1357.064825] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "refresh_cache-d8763396-efbb-4e45-b9c8-efdcd8be27e5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1357.115129] env[62552]: DEBUG nova.network.neutron [req-1240d292-ad14-4531-b25d-020b90f7a1b5 req-57fc8a33-4c7e-43f8-a1d8-58c9b2be0109 service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1357.183831] env[62552]: DEBUG nova.network.neutron [req-1240d292-ad14-4531-b25d-020b90f7a1b5 req-57fc8a33-4c7e-43f8-a1d8-58c9b2be0109 service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1357.687284] env[62552]: DEBUG oslo_concurrency.lockutils [req-1240d292-ad14-4531-b25d-020b90f7a1b5 req-57fc8a33-4c7e-43f8-a1d8-58c9b2be0109 service nova] Releasing lock "refresh_cache-d8763396-efbb-4e45-b9c8-efdcd8be27e5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1357.687720] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "refresh_cache-d8763396-efbb-4e45-b9c8-efdcd8be27e5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.687881] env[62552]: DEBUG nova.network.neutron [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Building network info cache for instance {{(pid=62552) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1358.219071] env[62552]: DEBUG nova.network.neutron [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Instance cache missing network info. {{(pid=62552) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1358.353324] env[62552]: DEBUG nova.network.neutron [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Updating instance_info_cache with network_info: [{"id": "68cc9439-cb9d-4f29-b1c7-048e9748adfd", "address": "fa:16:3e:9d:6f:a3", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68cc9439-cb", "ovs_interfaceid": "68cc9439-cb9d-4f29-b1c7-048e9748adfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1358.857072] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "refresh_cache-d8763396-efbb-4e45-b9c8-efdcd8be27e5" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1358.857393] env[62552]: DEBUG nova.compute.manager [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Instance network_info: |[{"id": "68cc9439-cb9d-4f29-b1c7-048e9748adfd", "address": "fa:16:3e:9d:6f:a3", "network": {"id": "87f72e5b-bfd0-4037-b6c5-132c5671d531", "bridge": "br-int", "label": "tempest-ServersTestJSON-884885644-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e4d140b840b453abcd828d2129de615", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68cc9439-cb", "ovs_interfaceid": "68cc9439-cb9d-4f29-b1c7-048e9748adfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62552) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1358.857849] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:6f:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '68cc9439-cb9d-4f29-b1c7-048e9748adfd', 'vif_model': 'vmxnet3'}] {{(pid=62552) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1358.865334] env[62552]: DEBUG oslo.service.loopingcall [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1358.865546] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Creating VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1358.865774] env[62552]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a3cbb3b-7f96-4627-8586-f493f41e70f7 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.885355] env[62552]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1358.885355] env[62552]: value = "task-1240301" [ 1358.885355] env[62552]: _type = "Task" [ 1358.885355] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.892403] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240301, 'name': CreateVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.395339] env[62552]: DEBUG oslo_vmware.api [-] Task: {'id': task-1240301, 'name': CreateVM_Task, 'duration_secs': 0.331751} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.395693] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Created VM on the ESX host {{(pid=62552) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1359.396222] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1359.396406] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1359.396725] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1359.396974] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49485ee4-0ed4-47ce-8e1d-6d131e7f0084 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.401332] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1359.401332] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dc657e-16a9-f0f2-fd0b-ef0ccc0e6375" [ 1359.401332] env[62552]: _type = "Task" [ 1359.401332] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.408260] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dc657e-16a9-f0f2-fd0b-ef0ccc0e6375, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.912259] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]52dc657e-16a9-f0f2-fd0b-ef0ccc0e6375, 'name': SearchDatastore_Task, 'duration_secs': 0.01156} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.912539] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1359.912775] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Processing image 562b48de-e081-4c62-a3e2-cf4ccfd0d1ee {{(pid=62552) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1359.913015] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1359.913176] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1359.913388] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1359.913649] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dbce6b7-669f-4eae-93c8-d17d90a93ef5 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.921602] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62552) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1359.921770] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62552) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1359.922436] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a913ed60-4e9e-403a-a09e-5245a5e3e781 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.928097] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1359.928097] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525085f8-d9fe-8ddd-0862-472ea2e70019" [ 1359.928097] env[62552]: _type = "Task" [ 1359.928097] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.934717] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525085f8-d9fe-8ddd-0862-472ea2e70019, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.438047] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]525085f8-d9fe-8ddd-0862-472ea2e70019, 'name': SearchDatastore_Task, 'duration_secs': 0.008608} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.438783] env[62552]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc7d4557-dbe8-4c6b-9660-0e976adab17b {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.443677] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1360.443677] env[62552]: value = "session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520db6b5-ffa0-5b32-ad9f-9e98e317bd5b" [ 1360.443677] env[62552]: _type = "Task" [ 1360.443677] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.450716] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520db6b5-ffa0-5b32-ad9f-9e98e317bd5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.954125] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': session[5228093d-40ea-0c0e-18b5-5ad5c7d92804]520db6b5-ffa0-5b32-ad9f-9e98e317bd5b, 'name': SearchDatastore_Task, 'duration_secs': 0.008973} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.954469] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk" {{(pid=62552) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1360.954785] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] d8763396-efbb-4e45-b9c8-efdcd8be27e5/d8763396-efbb-4e45-b9c8-efdcd8be27e5.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1360.955080] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b553bfe1-d2ec-4c75-a4e6-3bdb0ea77089 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.961111] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1360.961111] env[62552]: value = "task-1240302" [ 1360.961111] env[62552]: _type = "Task" [ 1360.961111] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.968779] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240302, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.471507] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240302, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.410295} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.473050] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee/562b48de-e081-4c62-a3e2-cf4ccfd0d1ee.vmdk to [datastore2] d8763396-efbb-4e45-b9c8-efdcd8be27e5/d8763396-efbb-4e45-b9c8-efdcd8be27e5.vmdk {{(pid=62552) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1361.473050] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Extending root virtual disk to 1048576 {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1361.473050] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8f3aeffa-7c70-457c-b316-1bb06c4b61a4 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.479233] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1361.479233] env[62552]: value = "task-1240303" [ 1361.479233] env[62552]: _type = "Task" [ 1361.479233] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.487013] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240303, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.988816] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240303, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056787} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.988901] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Extended root virtual disk {{(pid=62552) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1361.989586] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c22b3b0-1997-480c-bd73-55411f9a1574 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.010474] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] d8763396-efbb-4e45-b9c8-efdcd8be27e5/d8763396-efbb-4e45-b9c8-efdcd8be27e5.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1362.010691] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd95c49e-4a5e-40d1-bd18-bcc05d6994a3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.029468] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1362.029468] env[62552]: value = "task-1240304" [ 1362.029468] env[62552]: _type = "Task" [ 1362.029468] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.036559] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240304, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.538923] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240304, 'name': ReconfigVM_Task, 'duration_secs': 0.266852} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.539316] env[62552]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Reconfigured VM instance instance-00000075 to attach disk [datastore2] d8763396-efbb-4e45-b9c8-efdcd8be27e5/d8763396-efbb-4e45-b9c8-efdcd8be27e5.vmdk or device None with type sparse {{(pid=62552) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1362.539904] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f393f2d3-9fe1-4788-a06d-5f53a0212910 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.546250] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1362.546250] env[62552]: value = "task-1240305" [ 1362.546250] env[62552]: _type = "Task" [ 1362.546250] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.553719] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240305, 'name': Rename_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.056149] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240305, 'name': Rename_Task, 'duration_secs': 0.13017} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.056536] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Powering on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1363.056785] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1278dc76-8d65-489f-ad15-515635933ca8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.062370] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1363.062370] env[62552]: value = "task-1240306" [ 1363.062370] env[62552]: _type = "Task" [ 1363.062370] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.069363] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240306, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.573634] env[62552]: DEBUG oslo_vmware.api [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240306, 'name': PowerOnVM_Task, 'duration_secs': 0.410204} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.574054] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Powered on the VM {{(pid=62552) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1363.574054] env[62552]: INFO nova.compute.manager [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Took 7.76 seconds to spawn the instance on the hypervisor. [ 1363.574248] env[62552]: DEBUG nova.compute.manager [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1363.575012] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c516a7aa-f3fd-4c87-945b-c31f72260960 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.090493] env[62552]: INFO nova.compute.manager [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Took 12.42 seconds to build instance. [ 1364.592423] env[62552]: DEBUG oslo_concurrency.lockutils [None req-8d399cf0-b8b9-4388-be77-70d1b0d0072f tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.933s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1365.103227] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1365.103490] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1365.103677] env[62552]: DEBUG nova.compute.manager [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1365.104648] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce252fe-dd6a-4cb6-b315-2a833bd261d3 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.111711] env[62552]: DEBUG nova.compute.manager [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62552) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1365.112277] env[62552]: DEBUG nova.objects.instance [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lazy-loading 'flavor' on Instance uuid d8763396-efbb-4e45-b9c8-efdcd8be27e5 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1366.119204] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1366.119587] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c029dba6-70c3-48b8-b4ca-7f1832038584 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.126950] env[62552]: DEBUG oslo_vmware.api [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1366.126950] env[62552]: value = "task-1240307" [ 1366.126950] env[62552]: _type = "Task" [ 1366.126950] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.134728] env[62552]: DEBUG oslo_vmware.api [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240307, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.636841] env[62552]: DEBUG oslo_vmware.api [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240307, 'name': PowerOffVM_Task, 'duration_secs': 0.205658} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.637124] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1366.637338] env[62552]: DEBUG nova.compute.manager [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Checking state {{(pid=62552) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1366.638070] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f193403c-38e0-4c42-ae73-5933edc947d8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.148814] env[62552]: DEBUG oslo_concurrency.lockutils [None req-d00b9a7c-239a-478d-aec9-04d0e1743d14 tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.045s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1367.730051] env[62552]: DEBUG oslo_concurrency.lockutils [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1367.730352] env[62552]: DEBUG oslo_concurrency.lockutils [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1367.730570] env[62552]: DEBUG oslo_concurrency.lockutils [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1367.730762] env[62552]: DEBUG oslo_concurrency.lockutils [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1367.730938] env[62552]: DEBUG oslo_concurrency.lockutils [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1367.733131] env[62552]: INFO nova.compute.manager [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Terminating instance [ 1368.237136] env[62552]: DEBUG nova.compute.manager [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1368.237587] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1368.238308] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e83c4cd-34d7-421d-9f48-c9540836368f {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.246177] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1368.246421] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25d6ea44-75c2-442f-8b37-da4d3acd1c88 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.307289] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1368.307509] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1368.307685] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleting the datastore file [datastore2] d8763396-efbb-4e45-b9c8-efdcd8be27e5 {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1368.307929] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94817b87-09c5-4fcc-a6e7-023069149034 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.314174] env[62552]: DEBUG oslo_vmware.api [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1368.314174] env[62552]: value = "task-1240309" [ 1368.314174] env[62552]: _type = "Task" [ 1368.314174] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1368.321484] env[62552]: DEBUG oslo_vmware.api [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240309, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.823767] env[62552]: DEBUG oslo_vmware.api [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240309, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144072} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1368.824046] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1368.824248] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1368.824425] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1368.824595] env[62552]: INFO nova.compute.manager [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1368.824831] env[62552]: DEBUG oslo.service.loopingcall [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1368.825030] env[62552]: DEBUG nova.compute.manager [-] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1368.825128] env[62552]: DEBUG nova.network.neutron [-] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1369.047371] env[62552]: DEBUG nova.compute.manager [req-a7a0467d-ad52-4f8f-a853-94f9923850c3 req-c0bb6909-e41e-41ff-bcdd-df6d2552af2b service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Received event network-vif-deleted-68cc9439-cb9d-4f29-b1c7-048e9748adfd {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1369.047371] env[62552]: INFO nova.compute.manager [req-a7a0467d-ad52-4f8f-a853-94f9923850c3 req-c0bb6909-e41e-41ff-bcdd-df6d2552af2b service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Neutron deleted interface 68cc9439-cb9d-4f29-b1c7-048e9748adfd; detaching it from the instance and deleting it from the info cache [ 1369.047371] env[62552]: DEBUG nova.network.neutron [req-a7a0467d-ad52-4f8f-a853-94f9923850c3 req-c0bb6909-e41e-41ff-bcdd-df6d2552af2b service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1369.521603] env[62552]: DEBUG nova.network.neutron [-] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1369.549391] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66077dcc-4bd5-456c-b35d-f21b83db52ae {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.560173] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d24e19-9e93-410d-a7ca-52c5af7625eb {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.582099] env[62552]: DEBUG nova.compute.manager [req-a7a0467d-ad52-4f8f-a853-94f9923850c3 req-c0bb6909-e41e-41ff-bcdd-df6d2552af2b service nova] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Detach interface failed, port_id=68cc9439-cb9d-4f29-b1c7-048e9748adfd, reason: Instance d8763396-efbb-4e45-b9c8-efdcd8be27e5 could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1370.023740] env[62552]: INFO nova.compute.manager [-] [instance: d8763396-efbb-4e45-b9c8-efdcd8be27e5] Took 1.20 seconds to deallocate network for instance. [ 1370.529757] env[62552]: DEBUG oslo_concurrency.lockutils [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.530058] env[62552]: DEBUG oslo_concurrency.lockutils [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1370.530296] env[62552]: DEBUG nova.objects.instance [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lazy-loading 'resources' on Instance uuid d8763396-efbb-4e45-b9c8-efdcd8be27e5 {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1371.072078] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa3e658-9091-407c-95d3-bfbbca98839a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.079528] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581e101c-4115-4748-9fa9-59b65544566a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.109663] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989a07a4-d68b-4599-8453-fc26cd3a17cc {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.116419] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28dce9a-f5b7-40b5-8d5c-ab7d71c3b5c0 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.129008] env[62552]: DEBUG nova.compute.provider_tree [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1371.631719] env[62552]: DEBUG nova.scheduler.client.report [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1372.137083] env[62552]: DEBUG oslo_concurrency.lockutils [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.153995] env[62552]: INFO nova.scheduler.client.report [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted allocations for instance d8763396-efbb-4e45-b9c8-efdcd8be27e5 [ 1372.662438] env[62552]: DEBUG oslo_concurrency.lockutils [None req-850b0d13-b52c-49ba-a94f-d5a81abd86bb tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "d8763396-efbb-4e45-b9c8-efdcd8be27e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.932s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1373.441957] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "604e8472-7af4-49a1-9a0c-459482db9d6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1373.442207] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "604e8472-7af4-49a1-9a0c-459482db9d6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1373.442425] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "604e8472-7af4-49a1-9a0c-459482db9d6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1373.442616] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "604e8472-7af4-49a1-9a0c-459482db9d6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1373.442791] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "604e8472-7af4-49a1-9a0c-459482db9d6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1373.445161] env[62552]: INFO nova.compute.manager [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Terminating instance [ 1373.949545] env[62552]: DEBUG nova.compute.manager [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Start destroying the instance on the hypervisor. {{(pid=62552) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1373.949937] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Destroying instance {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1373.950726] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4cef63-e06d-4886-b79a-9caeb494ac3c {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.959691] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Powering off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1373.959916] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c695427-69be-4b0f-a9b0-d854f8f9805a {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.965676] env[62552]: DEBUG oslo_vmware.api [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1373.965676] env[62552]: value = "task-1240310" [ 1373.965676] env[62552]: _type = "Task" [ 1373.965676] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1373.973488] env[62552]: DEBUG oslo_vmware.api [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1374.475654] env[62552]: DEBUG oslo_vmware.api [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240310, 'name': PowerOffVM_Task, 'duration_secs': 0.178607} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1374.475918] env[62552]: DEBUG nova.virt.vmwareapi.vm_util [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Powered off the VM {{(pid=62552) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1374.476105] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Unregistering the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1374.476387] env[62552]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b88bec73-7e1c-4e68-9bbc-a1f74121b233 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.535901] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Unregistered the VM {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1374.536138] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Deleting contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1374.536354] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleting the datastore file [datastore2] 604e8472-7af4-49a1-9a0c-459482db9d6e {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1374.536628] env[62552]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-476899e9-8dfe-4045-ace7-c93cd16b5d46 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.542984] env[62552]: DEBUG oslo_vmware.api [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for the task: (returnval){ [ 1374.542984] env[62552]: value = "task-1240312" [ 1374.542984] env[62552]: _type = "Task" [ 1374.542984] env[62552]: } to complete. {{(pid=62552) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1374.550089] env[62552]: DEBUG oslo_vmware.api [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240312, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1375.052645] env[62552]: DEBUG oslo_vmware.api [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Task: {'id': task-1240312, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123891} completed successfully. {{(pid=62552) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1375.053020] env[62552]: DEBUG nova.virt.vmwareapi.ds_util [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted the datastore file {{(pid=62552) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1375.053103] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Deleted contents of the VM from datastore datastore2 {{(pid=62552) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1375.053287] env[62552]: DEBUG nova.virt.vmwareapi.vmops [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Instance destroyed {{(pid=62552) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1375.053463] env[62552]: INFO nova.compute.manager [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1375.053780] env[62552]: DEBUG oslo.service.loopingcall [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62552) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1375.053988] env[62552]: DEBUG nova.compute.manager [-] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Deallocating network for instance {{(pid=62552) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1375.054101] env[62552]: DEBUG nova.network.neutron [-] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] deallocate_for_instance() {{(pid=62552) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1375.550801] env[62552]: DEBUG nova.compute.manager [req-93fe866b-c51b-453c-87b3-8b9ba272d638 req-30627f68-672f-4fd4-8ed6-3d0b082c6548 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Received event network-vif-deleted-152d412c-2b92-4bd8-8182-8c8761fcdbfe {{(pid=62552) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1375.551075] env[62552]: INFO nova.compute.manager [req-93fe866b-c51b-453c-87b3-8b9ba272d638 req-30627f68-672f-4fd4-8ed6-3d0b082c6548 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Neutron deleted interface 152d412c-2b92-4bd8-8182-8c8761fcdbfe; detaching it from the instance and deleting it from the info cache [ 1375.551224] env[62552]: DEBUG nova.network.neutron [req-93fe866b-c51b-453c-87b3-8b9ba272d638 req-30627f68-672f-4fd4-8ed6-3d0b082c6548 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1376.024289] env[62552]: DEBUG nova.network.neutron [-] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Updating instance_info_cache with network_info: [] {{(pid=62552) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1376.054271] env[62552]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-899dfdf2-1c2a-4a1a-955a-95ce87293101 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.063754] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03aae1c-527a-48ae-a702-664e63497228 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.085502] env[62552]: DEBUG nova.compute.manager [req-93fe866b-c51b-453c-87b3-8b9ba272d638 req-30627f68-672f-4fd4-8ed6-3d0b082c6548 service nova] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Detach interface failed, port_id=152d412c-2b92-4bd8-8182-8c8761fcdbfe, reason: Instance 604e8472-7af4-49a1-9a0c-459482db9d6e could not be found. {{(pid=62552) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1376.527053] env[62552]: INFO nova.compute.manager [-] [instance: 604e8472-7af4-49a1-9a0c-459482db9d6e] Took 1.47 seconds to deallocate network for instance. [ 1377.033804] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1377.034103] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1377.034356] env[62552]: DEBUG nova.objects.instance [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lazy-loading 'resources' on Instance uuid 604e8472-7af4-49a1-9a0c-459482db9d6e {{(pid=62552) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1377.553254] env[62552]: DEBUG nova.scheduler.client.report [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Refreshing inventories for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1377.566089] env[62552]: DEBUG nova.scheduler.client.report [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Updating ProviderTree inventory for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1377.566330] env[62552]: DEBUG nova.compute.provider_tree [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Updating inventory in ProviderTree for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1377.576288] env[62552]: DEBUG nova.scheduler.client.report [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Refreshing aggregate associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, aggregates: None {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1377.592697] env[62552]: DEBUG nova.scheduler.client.report [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Refreshing trait associations for resource provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62552) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1377.616399] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27912b66-1a2c-4acd-8e12-16e5f05baf2e {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.623770] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ddd156-5646-4810-835d-d926cafbf7e8 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.653042] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524d3448-5df4-47f1-a445-7b74745f6eaa {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.659481] env[62552]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85685763-d6cc-48d3-bf54-99d175b49072 {{(pid=62552) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.671756] env[62552]: DEBUG nova.compute.provider_tree [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed in ProviderTree for provider: b52e1b51-8a99-4529-bad8-00cd27fc0fb8 {{(pid=62552) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1378.174507] env[62552]: DEBUG nova.scheduler.client.report [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Inventory has not changed for provider b52e1b51-8a99-4529-bad8-00cd27fc0fb8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62552) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1378.678942] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1378.697225] env[62552]: INFO nova.scheduler.client.report [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Deleted allocations for instance 604e8472-7af4-49a1-9a0c-459482db9d6e [ 1379.205401] env[62552]: DEBUG oslo_concurrency.lockutils [None req-12b410a8-00de-41ed-ac8d-62e577d475ee tempest-ServersTestJSON-1370032604 tempest-ServersTestJSON-1370032604-project-member] Lock "604e8472-7af4-49a1-9a0c-459482db9d6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.763s {{(pid=62552) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}